Professional Documents
Culture Documents
INTFtODw3TION
To perform repeatable tasks, robot manipulator
can be taught off-line by a so-called teaching and
playback scheme or can learn on-line from its
experience via a self learning control scheme without
human supervision.
Learning control theory has
attracted attention of control researchers since many
years [51 and recently has been developed for control
of robot manipulators [11-[41. The first learning
control scheme was proposed by Uchiyama [181 for
Motivated by the idea
control 1 ing mechanical arms.
that robots can learn autonomously from measurement
data of previous operations to improve their
performance, Arimoto and his research group [21
proposed a so-called Betterment Process whose
operation is based on a simple iteration rule. The
iteration rule autonomously generates a present
actuator input better than the previous one by using
stored data of errors in previous operation.
Application of the proposed learning control scheme
has been studied mostly for the problem of controlling
robot position. [ l ] and some attention was paid to
force* control [7]. Hara and others [61 considered
the synthesis of repetitive control systems for a
subclass of control systems whose outputs are
controlled to follow periodic reference commands.
Togai and Yamano 1171 studied the synthesis of a
discrete learning control scheme that relaxes the rank
condition in Arimoto learning control scheme. Craig
[41 studied an adaptive scheme for robot manipulators
for the case of repeated trials of a path. In [31, a
*)
In this paper, position implies both position
and orientation and force implies both force and
torque .
545
3.
;o(t) = JTz[Fd(t)
- Fc(t)l
(3)
(5)
(2).
546
(6)
R ( t ) i s positive d e f i n i t e f o r
learning scheme i s defined by
M(t);(t)
where
N ( t ) L ( t ) + G ( t ) o ( t ) + Td(t) = T
M ( t ) = M(qd)
(7)
(10)
qd
rd(t)
= M(qd)qd + N(q , q
d
and
o(t) = q(t)
The end-effector
T(t)
+ G(qd) + R(qd)
q,(t).
(11)
(13)
where
A (t) = K
A ( t ) + K A ( t ) + K A ( t ) + KfdA,,(t)(14)
w 3
A 4 ( t ) = J'SKJI
qd
(16)
qd
ukil(t)
= u(t)
( t )- Oil(t)o ( t ) - 0 i 2 ( t ) G k ( t )
M(t)ik(t)
27 1
~ ~ += ~
U ((t )t +) 0[6 ( 0 - 0 1 + a (Gd-Lk)J.
p
(17)
and
= uk(t)
(18)
(28)
o ( 0 ) = q ( 0 ) - q ( 0 )= 0 = wd(0)
(30)
(19)
REMARK:
I n the above proof, w e considered the special
case o f designing the l e a r n i n g control system based on
the pure p o s i t i o n c o n t r o l .
I n general, the above
proof can be extended t o the h y b r i d case where
concentration on c e r t a i n degrees o f freedom can be
s p e c i f i e d through the Selection Matrix S .
It i s
expected t h a t the h y b r i d learning control scheme
performs b e t t e r than the pure p o s i t i o n
learning
c o n t r o l scheme as we w i l l see when implementing the
h y b r i d learning control scheme i n the next section.
= u k ( t ) + OIAl(t)(od-uk)
+ [N(t) + @A2(t)lGk(t) +
+ [ G ( t ) + @Al(t)lok(t)
ukil(t)
= i ( 0 ) ;z ( 0 )= zd(0)
W
e note t h a t M ( t ) , N ( t ) + @ A 2 ( t ) = N ( t ) + a@, G ( t ) +
OAi(t) = G ( t ) + f3@ are continuously d i f f e r e n t i a b l e
matrices on
[O,Tl
because Xd(t.1 and F d ( t ) are
continuously d i f f e r e n t i a b l e on [O,Tl. I n a d d i t i o n ,
i, (0)= q (0) - q ( 0 ) = 0 = Ld (0)
(29)
where
(15)
pd3
J-l(I-S)JI
26 1
fP 4
pd3
A ( t ) = K A ( t ) + KfdA4(t)
A3(t) =
R ( t ) z k ( t ) + Q ( t ) t k (+t P
) (t)z (t)= vk(t)
-2
Main Theorem:
Consider the end-effector described by
(5) and adequately modelled by ( 7 ) .
I f the desired
(12)
= -A ( t ) o ( t ) - A 2 ( t ) L ( t ) + u ( t ) +
[O,TI.
ta(t
25 )
zk(0)
(9)
d'
#3(zd-zk)1
vkil(t)
(8)
N ( t ) = a/aq[n(q,q)iIq
= v (t) +
all
+ A 2 ( t ) ( L d - b k ) l . (23)
5.
547
5.1
End-Fffector Parapetem:
d = 29 in.,
1. = 11 in.,
[0;18 VAn.
Learning
0
O.8V/in.Ii
Control System
p o s i t i o n and force):
for
position,
0
.7Vsec/in.
=
(PO-control l e r s
for
both
a = af = 1/20; Bp = 6, = 19/20
.
+ G(q) + JT(q)Fc = T ( t )
:];
(331
[m,lm~1,-l2)/(3u)
(35)
G = [Gl G21'
where G G
1'
2'
1'
and
2
G, = {-mlg[2ul1 (; 1 , I m+i2i
,+21 2i
2 ) - i 2i
1 1+12)-1,u211/(4dl 1: 2u)
-mglmf21:ul(
G~
2 2
1 m
(37)
= 12-1:+
2
d2;
U2
2 m
1 ~ 2)1/ ( 4 d l 11;U)
= l:-l:+d2;
(38)
= [4d21:-uz11'2
(39)
6.
controlling
the
horizontal
position
x
of
the
end-effector and the v e r t i c a l contact f o r c e Fy the
end-effector applies on the r e a c t i o n table.
Thus for
t h i s p a r t i c u l a r case
the contact force vector i s
and the end-effector Jacobian
given by Fc= ( 0
i s given by
Fey?
5.2
EXPERIlIEHTuSTUDY
t h i s experimental
study, the end-eftector
as
shown i n Figure 3 is c o n t r o l l e d to follow
a desired
h o r i z o n t a l t r a j e c t o r y s p e c i f i e d by
x
= x
3 ( 3 e-l-tSt
1)
COXLUSION
I n t h i s paper,
we proposed a learning-based
h y b r i d c o n t r o l scheme f o r c o n t r o l l i n g p o s i t i o n / f o r c e
o f a CKCM end-effector.
The proposed control scheme
consists o f a h y b r i d c o n t r o l system serving as a
feedback c o n t r o l l e r and a learning c o n t r o l system t h a t
can
"learn"
to
improve
the
performance
of
of
theend-effector.
Employing
the
method
l i n e a r i z a t i o n around a desired force and p o s i t i o n
pattern, we showed t h a t the c o n t r o l l e r gains o f the
hybrid
control
system
can
be
selected
to
asymptotically s t a b i l i z e the end-effector.
After
t h a t . using the r e s u l t s i n 171. we proved t h a t the
c o n t r o l l e r gains o f the learnina c o n t r o l system can be
designed such t h a t the end-effector responses approach
t h e desired p a t t e r n uniformly as t h e number of t r i a l s
increases.
Experimental i n v e s t i g a t i o n performed on a
real-life
end-effector
showed t h a t
the proposed
c o n t r o l scheme provided excellent convergence o f the
e r r o r s i n force and postion.
I n order t o f u r t h e r
improve the end-ef f ector performance. it is suggested
t h a t f u t u r e research be extended t o studying
the
inplementation o f the proposed learning control scheme
i n a r o b o t i c system whose feedback control system i s
h @ r i d and adaptive [131, [161.
548
ll
7
- -
9.
Nguyen, C.C., "Arbitrary Eigenvalue Assignments
for
Linear
Time-Varying
Multivariable
Controi
Systems, International Journal of Control, Vol. 45,
NO. 3, pp. 1051-1057, 1987.
10. Nguyen, C.C., Pooran, F.J., and Premack, T.,
"Control of Robot Manipulator Compliance, " in Recent
Trends in Robotics: Modeling, Control and Education,
edited by M. Jamshidi, J.Y.S. Luh, and M . Shahinpoor,
North Holland. New York, pp. 237-242, 1986.
11. Nguyen, C. C., Pooran. F. J. and Premack, T.,
"Modified Hybrid Control of Robot Manipulators for
ISMM
High Precision Assembly Operations." h o c .
Conference
on Computer Applications
in Design,
Simulation
and Analysis, Honolulu, Hawaii, pp.
191-195, Feb. 1988.
12. Nguyen, C. C., and Pooran, F. J., "Kinematics and
Dynamics of a Six-Degree-of-Freedom Robot Manipulator
with Closed-Kinematic Chain Mechanism." Proc. Second
Int. Symp. Robotics and Manufact. Research, Education
and Applications, New Mexico, November 1988.
13. Nguyen, C.C. and Pooran, F.J., "Adaptive Control
of Robot Manipulators with Closed-Kinematic Chain
Hechanism." h o c . Second International Symposium on
Robotics and Manufacturing Research, Education and
Applications, New Mexico, November 1988.
14. Premack, T., et.al., "Design and Implementation
of a Compliant Robot with Force Feedback and Strategy
Planning Sofuare," NASA Technical Memorandum 86111,
1984.
15. Raibert, M.
H.,
and Craig, J.J., "Hybrid
Positioflorce Control of Manipulators." ASME, Journal
of Dynamic Systems, Measurement and Control, 102, pp.
126-133, June 1981.
16. Seraji, H., "Adaptive Force and Position Control
of Manipulators." Journal of Robotics Systems, 4(4),
pp. 551-578, 1987.
17. Togai, M . , and Yamano, O., "Analysis and Design
of an Optimal Learning Control Scheme for Industrial
Robots: A Discrete System Approach", Proc. 24th Conf.
on Decision and Control, pp. 1399-1404, Dec. 1985.
18. Uchiyama, M . , "Formation of High Speed Motion
Pattern of Mechanical Arm by Trial". Transactions.
Society of Instrument and Control Engineers, Vol. 19.
(I 5, pp. 706-712, 1978.
ACKNOWLEDCMENTS
The research work presented in this paper has
been sponsored by Goddard Space Flight Center (NASA)
under the research grant with Grant Number NAG 5-780.
The authors would like to express their appreciation
to NASA for continuous support of the research
project.
7.
I'
REFERENCES
1.
Arimoto, S., et al., "Can Robot Learn by
Themselves?", Robotic Research, 3rd Int. Symposium, H.
Hanafusa and H, Inoue, E&, MIT Press, Massachusetts,
1985.
2.
Arimoto, S. et al., "Bettering Operation of
Dynamic S y s t e m by Learning: A New Control Theory for
Servomechanism or Mechatronics System", h o c . 23rd
Conf. Decision and Control, pp. 1064-1069, December
1984.
3.
Atkeson, C. G.,
and McIntyre, J., "Robot
Trajectory Learning Through Practice", h o c . IEEE
Conf. Robotics and Automation, pp. 1737-1742, San
Francisco, 1986.
4.
Craig, J. J., "Adaptive Control of Manipulators
Through Repeated Trials", h o c . American Control
Conf., pp 1566-1573, June 1984.
5.
Fu, K.S., "Learning Control Systems-Review and
Outlook", IEEE Trans. Automatic Control, AC-15, pp.
210-221, April 1970.
6. Hara, S., Omata, T., and Nakano, M . . "Synthesis
of Repetitive Control Systems and Its Applications",
R o c . 24th Conf. on Decision and Control, pp.
1387-1392, Dec. 1985.
7.
Kawamura, S., Miyazaki, F . , and Arimoto, S.,
"Applications of Learning Method for Dynamic Control
of Robot Manipulators", Proc. 24th Conf. on Decision
and Control, pp. 1381-1386, Dec. 1985.
8.
Mita, T. and Kato, E., "Iterative Control and Its
Application to Motion Control of Robot Arm: A Direct
Approach to Servo Problems", h o c . 24th Conf. Decision
and Control, pp. 1393-1398, Dec 1985.
FORYARO
KINEMATIC
TRANSFORMA
TlON
POSlTlON
SENSORS
b
END-EFFECTOR
AND
ENVIRONMENT
FC
CF-1-J
FORCE
~-
COORDINATE
TRANSFORMA
FORCE
SENSORS
TION
Fig. 2:
I,
'actuator' .
o rm
F i g . 1:
4e
"
'k
'kI1
L S I RAM
Fig. 4:
0.5
0.5
k=2
-0.5
-0.5
0
10
15
20
Time (sec)
10
20
15
50
-.
5
Time (sec)
Fig. 5:
10
Time (sec)
3-
2-
2-
2 - .
1-
0-
-1
k=l
1-
0-
-1
k=2
1--
k=10
0-
-1
15
20