impedance adaptation for optimal robot–environment...

18
Impedance adaptation for optimal robot–environment interaction Article (Accepted Version) http://sro.sussex.ac.uk Ge, Shuzhi Sam, Li, Yanan and Wang, Chen (2013) Impedance adaptation for optimal robot– environment interaction. International Journal of Control, 87 (2). pp. 249-263. ISSN 0020-7179 This version is available from Sussex Research Online: http://sro.sussex.ac.uk/id/eprint/72085/ This document is made available in accordance with publisher policies and may differ from the published version or from the version of record. If you wish to cite this item you are advised to consult the publisher’s version. Please see the URL above for details on accessing the published version. Copyright and reuse: Sussex Research Online is a digital repository of the research output of the University. Copyright and all moral rights to the version of the paper presented here belong to the individual author(s) and/or other copyright owners. To the extent reasonable and practicable, the material made available in SRO has been checked for eligibility before being made available. Copies of full text items generally can be reproduced, displayed or performed and given to third parties in any format or medium for personal research or study, educational, or not-for-profit purposes without prior permission or charge, provided that the authors, title and full bibliographic details are credited, a hyperlink and/or URL is given for the original metadata page and the content is not changed in any way.

Upload: others

Post on 27-Mar-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

Impedance adaptation for optimal robot–environment interaction

Article (Accepted Version)

http://sro.sussex.ac.uk

Ge, Shuzhi Sam, Li, Yanan and Wang, Chen (2013) Impedance adaptation for optimal robot–environment interaction. International Journal of Control, 87 (2). pp. 249-263. ISSN 0020-7179

This version is available from Sussex Research Online: http://sro.sussex.ac.uk/id/eprint/72085/

This document is made available in accordance with publisher policies and may differ from the published version or from the version of record. If you wish to cite this item you are advised to consult the publisher’s version. Please see the URL above for details on accessing the published version.

Copyright and reuse: Sussex Research Online is a digital repository of the research output of the University.

Copyright and all moral rights to the version of the paper presented here belong to the individual author(s) and/or other copyright owners. To the extent reasonable and practicable, the material made available in SRO has been checked for eligibility before being made available.

Copies of full text items generally can be reproduced, displayed or performed and given to third parties in any format or medium for personal research or study, educational, or not-for-profit purposes without prior permission or charge, provided that the authors, title and full bibliographic details are credited, a hyperlink and/or URL is given for the original metadata page and the content is not changed in any way.

Page 2: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of ControlVol. 00, No. 00, Month 201x, 1–17

RESEARCH ARTICLE

Impedance Adaptation for Optimal Robot-Environment Interaction

Shuzhi Sam Gea,c∗, Yanan Lia,b and Chen Wanga

aSocial Robotics Laboratory, Interactive Digital Media Institute, and Department of Electrical and ComputerEngineering, National University of Singapore, Singapore119077

bNUS Graduate School for Integrative Sciences and Engineering, National University of Singapore, Singapore119613

cRobotics Institute, and School of Computer Science and Engineering, University of Electronic Science andTechnology of China, Chengdu 610054, China

(Received 00 Month 201x; final version received 00 Month 201x)

In this paper, impedance adaptation is investigated for robots interacting with unknown environments. Impedance control is em-ployed for the physical interaction between robots and environments, subject to unknown and uncertain environments dynamics.The unknown environments are described as linear systems with unknown dynamics, based on which the desired impedancemodel is obtained. A cost function that measures the tracking error and interaction force is defined, and the critical impedanceparameters are found to minimize it. Without requiring the information of the environments dynamics, the proposed impedanceadaptation is feasible in a large number of applications where robots physically interact with unknown environments. The validityof the proposed method is verified through simulation studies.

Keywords: robot-environment interaction; impedance adaptation; optimal control

1 Introduction

In social applications such as elderly care, health care andhuman-robot collaboration, environmentsare typically unknown to robots and there exist uncertainties due to many factors. Therefore, controlof interaction between robots and environments is essential and there has been much effort made onthis topic. In the literature of interaction control, two approaches are widely studied, i.e., hybrid posi-tion/force control Craig and Raibert (1979), Huang et al. (2006), Li et al. (2007) and impedance controlHogan (1985). Compared to hybrid position/force control, impedance control is more feasible in thesense that it does not require the decomposition of two directions. Besides, impedance control is pre-ferred for its better robustness Colgate and Hogan (1988).

Under impedance control, robots are governed to be compliant to the interaction force exerted by envi-ronments and thus the safety of both robots and environmentscan be guaranteed. Specifically, imposinga passive impedance model to robots will guarantee the interaction stability if environments are alsopassive Colgate and Hogan (1988). In the early research works of impedance control, a desired passiveimpedance model is usually prescribed and then the effort isfocused on handling the uncertainties inrobots dynamics. These works include adaptive impedance control such as Colbaugh et al. (1991), Luand Meng (1991) and learning impedance control such as Cheahand Wang (1998), Wang and Cheah(1998), Li et al. (2012). However, in many situations, to impose a passive impedance model to the robotis too conservative, and the environments dynamics can be taken into consideration to obtain desiredimpedance model Buerger and Hogan (2007). Besides, a fixed prescribed impedance model does not

∗Corresponding author. Tel: (+65) 6516 6821, Fax: (+65) 67791103, E-mail: [email protected]

ISSN: 0020-7179 print/ISSN 1366-5820 onlinec© 201x Taylor & Francis

DOI: 10.1080/0020717YYxxxxxxxxhttp://www.informaworld.com

Page 3: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

2 Shuzhi Sam Ge, Yanan Li and Chen Wang

suffice in many applications. For example, variable impedance control is necessary in human-robot col-laboration Tsumugiwa et al. (2001, 2002) and explosive movement Braun et al. (2012a,b). Althoughthe methods discussed in Tsumugiwa et al. (2001, 2002) provide a better control performance in thesense of more efficient human-robot collaboration, the resulted impedance parameters (mass, dampingand stiffness) are obtained in a heuristical way and cannot be easily extended to other applications. Tocope with this problem, iterative learning has been studiedto obtain impedance parameters subject tounknown environments in an analytic way. It has been generally acknowledged that such an ability toimprove performance by repeating a task is an important control strategy of the human being De Rooveret al. (2000), Xu et al. (2000). Pioneered by Arimoto et al. (1984a,b), iterative learning control has beenwidely investigated for robot control. In Cohen and Flash (1991), associative search network is adoptedfor the impedance learning and the resulted impedance control is applied to a wall-following task. InYang and Asada (1996), an internal model based impedance learning method is developed and used in ahigh-speed insertion application. In Tsuji and Morasso (1996), neural networks are employed to updateboth the impedance parameters and rest position iteratively. Compared to iterative impedance learningdiscussed above, impedance adaptation is more interestingyet it is more challenging. It is interesting be-cause it does not require the robot to repeat operations to learn the desired impedance parameters. Thisis important because to make the robot repeat operations maycause inconvenience in many situations. Itis challenging because to develop an adaptive scheme usually requires that a certain variable is invariantbut this is difficult to satisfy in the case of dynamically changing environment. There has been researcheffort on impedance adaptation in the literature, althoughit is less compared to that on impedance learn-ing. In Uemura and Kawamura (2009), stiffness is updated to minimize the actuator torque by takingresonance into consideration. In Stanisic and Fernandez (2012), the switching strategies of impedanceparameters are discussed in order to dissipate the system energy and realize a “soft” interaction.

In the development of impedance learning and adaptation, optimization plays an important role be-cause the control objective of impedance control includes both the force regulation and trajectory track-ing and usually it is the compromise of these two objectives.In the case of impedance adaptation, acost function or a reward function is defined to describe the interaction performance, and impedanceparameters are expected to be adjusted to minimize the cost function or maximize the reward function.In Johansson and Spong (1994), the well-known linear quadratic regulator (LQR) is utilized to deter-mine the desired impedance parameters with the environmentdynamics knowna priori. In Matinfar andHashtrudi-Zaad (2005), impedance parameters are adjustedas the online solutions of the defined LQRproblem, instead of fixing the impedance parameters obtained based on LQR as in Johansson and Spong(1994), However, the environment dynamics are also assumedto be known in Matinfar and Hashtrudi-Zaad (2005). Recalling LQR in Kirk (1970), it is difficult to find the solution of the Riccati equationif the linear system under study is unknown. Therefore, whenthe system dynamics are unknown, themethods proposed in Johansson and Spong (1994), Matinfar and Hashtrudi-Zaad (2005) are not applica-ble. To solve the optimal control problems in the case of unknown system dynamics, adaptive dynamicprogramming (ADP) or reinforcement learning (RL) has been widely studied in the literature Werbos(1990a,b, 1992), Bertsekas (1995), Werbos (2009), Lewis and Vrabie (2009), Wang et al. (2009). ADPis constructed based on the idea of how biological system interacts with the surrounding environment.In the scheme of ADP, the control system is defined as an actor or agent, modifying its action basedon the feedback information of the environment. The actor oragent is rewarded or punished for a con-trol action which is evaluated by a critic Lewis and Vrabie (2009), Wang et al. (2009). Among all ADPapproaches, most recognized discrete ADP algorithms are the heuristic dynamic programming (HDP),action-dependent heuristic dynamic programming (ADHDP) or Q-learning Watkins (1989), globalizedDHP (GDHP) and dual-heuristic programming (DHP). The common feature of these ADP algorithmsis that the design of optimal controller only requires partial information of the system model to be con-trolled. There are existing works where ADP is adopted for the impedance adaptation of robot armcontrol. In Kim et al. (2010), natural actor-critic algorithm is adopted and the damping and stiffness ma-trices are updated according to defined reward functions. InBuchli et al. (2011), the policy improvementwith path integrals (PI2) algorithm is integrated with the reinforcement learning algorithm to achievevariable impedance control. However, as in Arimoto et al. (1984a,b), Cohen and Flash (1991), Yang and

Page 4: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 3

Asada (1996), Tsuji and Morasso (1996), a learning process is still required in Kim et al. (2010), Buchliet al. (2011) for the robot to repeat operations to learn the desired impedance parameters. To solve thisproblem, this paper aims to develop impedance adaptation inthe case of unknown environments dynam-ics. The method to be developed is based on the latest result of ADP in Jiang and Jiang (2012), wherethe solution of adaptive optimal control is obtained subject to unknown system dynamics. Two generalmodels of environments are considered, one of which includes damping and stiffness, and the other oneincludes mass, damping and stiffness. These two models are described as linear systems with unknowndynamics. While ADP in Jiang and Jiang (2012) is only for the state regulation, it is further modified tohandle the trajectory tracking. The developed impedance adaptation will result in the desired impedanceparameters that are able to guarantee the optimal interaction, subject to unknown environments.

Based on the above discussion, we highlight the contributions of this paper as follows:

(i) Environment dynamics are taken into consideration in the analysis of optimal robot-environment in-teraction, and they are described as linear systems with unknown dynamics.

(ii) ADP for systems with unknown dynamics is modified such that trajectory tracking is achievable andthe desired impedance model can be obtained.

(iii) Impedance parameters of robots are obtained subject to unknown environments, which guarantee theoptimal robot-environment interaction in the sense of trajectory tracking and force regulation.

The rest of the paper is organized as follows. In Section 2, the dynamics of the robot and environ-ment are described, and impedance control and the objectiveof this paper are discussed. In Section 3,impedance adaptation is developed for two general kinds of environments, such that the optimal interac-tion is achieved subject to unknown environments. In Section 4, the validity of the proposed method isverified through simulation studies. Section 5 concludes this paper.

2 Problem Formulation

2.1 System Description

The system under study includes a rigid robot arm and an environment, where the end-effector of therobot arm physically interacts with the environment. Thereis a force sensor at the end-effector of therobot arm which measures the interaction force between the robot arm and the environment, as shown inFig. 1.

Consider the robot kinematics as below

x(t) = φ(q) (1)

wherex(t) ∈ Rn andq ∈ R

n are positions/oritations in the Cartesian space and joint coordinates in thejoint space, respectively. Differentiating (1) with respect to time results in

x(t) = J(q)q (2)

whereJ(q) ∈ Rn×n is the Jacobian matrix.

The robot dynamics in the joint space are given by

M(q)q + C(q, q)q +G(q) = τ + JT (q)f (3)

whereM(q) ∈ Rn×n is the inertia matrix;C(q, q)q ∈ R

n denotes the Coriolis and Centrifugal force;G(q) ∈ R

n is the gravitational force;τ ∈ Rn is the vector of the control input; andf(t) ∈ R

n denotesthe force exerted by the environment.

The other part of the system under study is the environment. Without loss of generality, two kindsof environments are considered in this paper, of which the dynamics are respectively described by the

Page 5: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

4 Shuzhi Sam Ge, Yanan Li and Chen Wang

ME

CE

GE

Y

X

(a)

CE

GE

Y

X

(b)

Figure 1.: System under study: (a) mass-damping-stiffnessenvironment, and (b) damping-stiffness envi-ronment

following models

ME x+ CEx+GEx = −f (4)

CE x+GEx = −f (5)

whereME , CE andGE are unknown mass, damping and stiffness matrices of the environment models,respectively. Compared to the second model (5), there is a mass matrix in the first model (4). It will beshown that different impedance models of the robot arm are required for different environments (4) and(5) to achieve the optimal interaction.

Remark 1 : The above two kinds of environments represent a large range of environments. For exam-ples, model (4) may describe the dynamics of human limb in physical human-robot interaction Rahmanet al. (2002), and model (5) may represent the viscoelastic object in robotic manipulation.

Remark 2 : ME , CE andGE are assumed to be unknown constant matrices in this paper. While it isvalid in many applications, these matrices can be time-varying in some other applications. The latterassumption makes the problem studied in this paper more complicated and it will be further investigatedin the future work.

2.2 Impedance Control

As discussed in the Introduction, impedance control is employed for robots interacting with environ-ments. To implement impedance control, a two-loop control framework is usually adopted, as shown inFig. 2.

In this framework, the outer-loop is dedicated to generate the virtual desired trajectory in the jointspace, i.e.,qd. In particular, the desired impedance model in the Cartesian space is given by

f = Z(xd, x0) (6)

wherexd is the desired trajectory,x0 is the virtual desired trajectory in the Cartesian space, and Z(·)is a target impedance function to be determined. Then, the virtual desired trajectory in the joint space

Page 6: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 5

Robot

Arm

+ Position

Control

f

q

Impedance

Model

qd

-

Inverse

Kinematics

x0xd

Figure 2.: Impedance Control Diagram

qd =∫ t

0 J−1(q(v))x0(v)dv is obtained according to the interaction forcef and the impedance model

(6).

Remark 3 : Model (6) is a general impedance model which defines the impedance relationship betweeninteraction force and position. In practical implementations, a typical impedance model is usually givenby Mdx0 + Cdx0 + Gd(x0 − xd) = −f , whereMd, Cd andGd are desired inertial, damping andstiffness matrices, respectively. A more simplified model,e.g., stiffness modelGd(x0 − xd) = −f , maybe adopted in a specific situation.

The inner-loop is to guarantee the trajectory tracking, i.e., limt→∞ q(t) = qd(t). Trajectory trackingof a robot arm has been extensively studied in the literatureGe et al. (1998), and will not be discussed inthis paper. For the analysis simplicity, it is assumed that there is an ideal inner-loop position controllersuch thatq(t) = qd(t) and thusx(t) = x0(t). In this way, the desired impedance model becomes

f = Z(xd, x) (7)

In (7), the impedance functionZ(·) is determined such that a certain interaction requirement is satisfied.This is non-trivial considering that the environment dynamics (4) and (5) are unknown. As discussed inthe Introduction, iterative learning has been studied to cope with this problem, which however requiresrepetitive motions. In this paper, we aim to achieve the sameobjective while avoiding the learningprocess. This is the motivation to develop impedance adaptation in the rest of this paper.

2.3 Preliminary: Adaptive Optimal Control

The adaptive optimal control proposed in Jiang and Jiang (2012) is briefly introduced in this subsec-tion, of which the results will be used for the development ofthe impedance adaptation.

Consider the following linear system

ξ = Aξ +Bu (8)

whereξ ∈ Rp is the system state,u ∈ R

r is the system input, andA ∈ Rp×p andB ∈ R

p×r are unknownconstant matrices.

The following system input

u = −Kkξ (9)

minimizes a defined cost function

Γ =

0[ξTQξ + uTRu]dt (10)

whereQ ∈ Rp×p andR ∈ R

r×r are the weights of the state and the input which satisfyQ = QT ≥ 0andR = RT > 0, andKk with the iteration numberk is a matrix obtained by following the proceduresas below:

Page 7: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

6 Shuzhi Sam Ge, Yanan Li and Chen Wang

Step 1: Employu = K0ξ+ ν as the input on the time interval[t0, tl], whereK0 stabilizes the system (8)andν is the exploration noise to satisfy the persistent excitation (PE) condition. Computeδξ, Iξ andIuuntil the following rank condition is satisfied

rank([Iξ, Iu]) =p(p+ 1)

2+ pr (11)

Step 2: SolvePk andKk+1 according to

[

Pk

vec(Kk+1)

]

= (ΘTkΘk)

−1ΘTk Ξk (12)

Step 3: Let k+1 → k and repeat Step 2 untilPk −Pk−1 ≤ ǫ whereǫ > 0 is a predefined constant. AndKk in (9) is obtained.

In Step 1, the following definitions are needed:

ξ = [ξ21 , ξ1ξ2, . . . , ξ1ξp, ξ22 , ξ2ξ3, . . . , ξp−1ξp, ξ

2p]

T

δξ = [ξ(t1)− ξ(t0), ξ(t2)− ξ(t1), . . . , ξ(tl)− ξ(tl−1)]T

Iξ = [

∫ t1

t0

ξ ⊗ ξdt,

∫ t2

t1

ξ ⊗ ξdt, . . . ,

∫ tl

tl−1

ξ ⊗ ξdt]T

Iu = [

∫ t1

t0

ξ ⊗ udt,

∫ t2

t1

ξ ⊗ udt, . . . ,

∫ tl

tl−1

ξ ⊗ udt]T (13)

whereξi, i = 1, . . . , p are the elements ofξ, l is a positive integer, and “⊗” is the Kronecker product.In Step 2, the following definitions are needed:

Pk = [P1,1, 2P1,2, . . . , 2P1,p, P2,2, 2P2,3, . . . , 2Pp−1,p, Pp,p]T

Θk = [δξ ,−2Iξ(Ip ⊗KTk R)− 2Iu(Ip ⊗R)]

Ξk = −Iξvec(Qk) (14)

wherePi,j , i = 1, . . . , p, j = 1, . . . , p are the elements ofPk, Ip is ap−dimentional unit matrix, “vec”is the operator that “stretches” a matrix to a vector, andQk = Q+KT

k RKk.

Remark 4 : Note thatA andB in (8) are not used in the above procedures so the resulted adaptiveoptimal control is applicable to the system with unknown dynamics. This is a favorable property com-pared to the traditional LQR since in many situations the system dynamics are difficult to obtain, if notimpossible.

3 Impedance Adaptation

This section is dedicated to develop impedance adaptation to achieve optimal interaction subject tounknown environments. Two kinds of environments describedby (4) and (5) will be considered in twosperate subsections, respectively.

3.1 Mass-Damping-Stiffness Environment

The mass-damping-stiffness environment described by (4) is considered in this subsection. By tak-ing the environment dynamics into consideration, we will develop impedance adaptation such that the

Page 8: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 7

following cost function is minimized

Γ =

0[xTQ1x+ (x− xd)

TQ2(x− xd) + fTRf ]dt (15)

whereQ1 ∈ Rn×n andQ2 ∈ R

n×n are the weights of the velocity and the trajectory tracking error,respectively, andR ∈ R

n×n has been defined in (8) and is the weight of the interaction force. Besides,Q1 = QT

1 ≥ 0 andQ2 = QT2 ≥ 0.

Remark 5 : Cost functions similar to (15) have been discussed in the related works Johansson andSpong (1994), Matinfar and Hashtrudi-Zaad (2005), which represent the compromise/combination ofthe force regulation and trajectory tracking and determinethe interaction performance. Note that thesecost functions are different from that in the traditional LQR problem, where the cost function usuallyincludes the control input and trajectory tracking error. In the traditional LQR problem, the system understudy is the robot itself, while the interaction system under study in this paper includes both the robotand the environment.

Comparing the cost function for a general linear system (10)and the defined cost function in this paper(15), some manipulations are needed to make them identical.In particular, we consider

ξ = [xT , xT , zT ]T (16)

wherez ∈ Rm is the state of the following system

{

z = Uz

xd = V z(17)

whereU ∈ Rm×m andV ∈ R

n×m are two known matrices.

Remark 6 : The linear system (17) is to determine the desired trajectory xd and provides the feasibilityto employ the optimal control in trajectory tracking problem. (17) is able to generate a large variety ofdesired trajectories, e.g., polynomial functions of time of any order.

Then, according to (15) and (17), we have

Γ =

0(xTQ1x+ [xT xTd ]

[

Q2 −Q2

−Q2 Q2

] [

x

xd

]

+ fTRf)dt

=

0(xTQ1x+ [xT zT ]

[

Q2 −Q2V

−V TQ2 VTQ2V

] [

x

z

]

+ fTRf)dt

=

0(ξTQξ + fTRf)dt (18)

whereQ =

[

Q1 00 Q′

2

]

with Q′

2 =

[

Q2 −Q2V

−V TQ2 VTQ2V

]

.

Considering the defined state (16), we rewrite (4) in the following state-space form

ξ = Aξ +Bf (19)

whereA =

−M−1E CE −M−1

E GE 0In 0 00 0 U

andB =

−M−1E

00

. Note thatA andB include the environ-

ment dynamics and they are unknown. This is the main reason tocause the difficulty of determining thedesired impedance functionZ(·) in (7).

Page 9: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

8 Shuzhi Sam Ge, Yanan Li and Chen Wang

If we take the interaction forcef in (19) as the “system input” to the environment dynamics, itcan beobtained as follows such that the cost function (15) is minimized

f = −Kkξ (20)

whereKk is obtained according to the procedures described in Section 2.3.

Remark 7 : For the augumented system, bothx andx are controllable while the uncontrollable statev

is asymptotically stable, so the system is stabilizable.To understand (20) in the sense of impedance control, we assume that the optimal control has been

obtained, which is

f = −Kξ = −R−1BTPξ (21)

whereK is the optimal feedback gain matrix andP = P T ∈ R(2n+m)×(2n+m) is the solution of the

following Riccati equation

PA+ATP − PBR−1BTP +Q = 0 (22)

Denote

P =

P1 P2 P3

∗ ∗ ∗∗ ∗ ∗

(23)

whereP1 ∈ Rn×n, P2 ∈ R

n×n, P3 ∈ Rn×m and “∗” stand for elements which we do not care about.

Substituting (23) into (21) leads to

f = −R−1P1x−R−1P2x−R−1P3z

= −R−1P1x−R−1P2x−R−1P4xd (24)

whereP4 = P3(VTV )−1V T .

Comparing the above equation (24) with the desired impedance model (7), the exact impedance func-tion which guarantees the optimal interaction is obtained.Recalling the implementation of impedancecontrol as described in Section 2.2, the virtual desired trajectory in the Cartesian spacex is obtainedaccording to (24) with measuredf and givenxd, and the inner-position control loop is to guaranteethe trajectory tracking in the joint space. In this way, the optimal interaction is achieved and (24) is theresulted impedance function in the presence of unknown environment dynamics.

3.2 Damping-Stiffness Environment

For the environment without mass, i.e., the damping-stiffness environment described by (5), the re-sulted impedance adaptation is different and it is discussed in this subsection.

First, we consider to minimize the following cost function

Γ′ =

0[(x− xd)

TQ2(x− xd) + fTRf ]dt (25)

Remark 8 : Note that the component to penalize the velocity in (15) has disappeared in (25). Thereason is that the velocity is not the state of the environment (5), which will be further explained in thefollowing.

Page 10: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 9

Correspondingly, we consider the stateξ′ = [xT , zT ]T , wherez ∈ Rm has the same meaning as

defined in (17). Based on the similar manipulation as in the previous section, (25) can be rewritten as

Γ′ =

0(ξ

′TQξ′ + fTRf)dt (26)

whereQ′ =

[

Q2 −Q2V

−V TQ2 VTQ2V

]

.

Similarly, (5) can be rewritten in the following state-space form

ξ′ = A′ξ′ +B′f (27)

whereA′ =

[

−C−1E GE 00 U

]

andB′ =

[

−C−1E

0

]

.

If we take the interaction forcef in (27) as the “system input” to the environment dynamics (5), it canbe obtained as follows such that the cost function (25) is minimized

f = −K ′

kξ′ (28)

whereK ′

k is obtained according to the procedures described in Section 2.3.To understand (28) in the sense of impedance control, we assume that the optimal control has been

obtained, which is

f = −K ′ξ′ = −R−1B′TP ′ξ′ (29)

whereK ′ is the optimal feedback gain matrix andP ′ = P′T ∈ R

(n+m)×(n+m) is the solution of thefollowing Riccati equation

P ′A′ +A′TP ′ − P ′B′R−1B

′TP ′ +Q′ = 0 (30)

Denote

P ′ =

[

P ′

1 P′

2∗ ∗

]

(31)

whereP ′

1 ∈ Rn×n andP ′

2 ∈ Rn×m. Substituting (31) into (29) leads to

f = −R−1P ′

1x−R−1P ′

2z

= −R−1P ′

1x−R−1P ′

3xd (32)

whereP ′

3 = P ′

2(VTV )−1V T .

It is found that the resulted control is variable stiffness control, i.e., there is no damping componentas in (24). In this sense, we conclude that both the damping and stiffness components are needed inthe impedance adaptation for the optimal interaction with the mass-damping-stiffness environments, andonly the stiffness component is needed for the optimal interaction with the damping-stiffness environ-ments. Similarly as in Section 3.1, the desired impedance functionZ(·) in (7) is obtained as (32), whichguarantees the optimal interaction subject to unknown environment dynamics (5).

Remark 9 : The adaptive optimal control proposed in Jiang and Jiang (2012) is limited to linear andtime invariant LQR problem, so the methods developed in Sections 3.1 and 3.2 are also limited to linearand time invariant environment dynamics. However, these methods could also be further modified to suitthe case where the parameters of the environment dynamics are changing slowly, using a time varyingparameter estimation.

Page 11: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

10 Shuzhi Sam Ge, Yanan Li and Chen Wang

4 Simulation Study

4.1 Simulation Conditions

In this section, we consider a robot arm with two revolute joints physically interacting with two envi-ronments, as discussed through this paper and shown in Figs.1(a) and 1(b). The simulation is conductedwith the Robotics Toolbox Corke (1996).

The parameters of the robot arm are:m1 = m2 = 2.0kg, l1 = l2 = 0.2m, i1 = i2 = 0.027kgm2,lc1 = lc2 = 0.1m, wheremj , lj , ij , lcj , j = 1, 2, represent the mass, the length, the inertia about thez-axis that comes out of the page passing through the center of mass, and the distance from the previousjoint to the center of mass of linki, respectively.

The initial joint coordinates of the robot arm areq1 = −π3 andq2 = 2π

3 , and thus the initial positionin the Cartesian space isxd = [0.2 0]T . The desired trajectory in the Cartesian space is determined by(17) with U = 1 andV = 0.3. It is assumed that the force exerted by the environment is only in X

direction and thus the robot arm inY direction is interaction-free. Nevertheless, the inner-loop positioncontrol is designed for both joints. In particular, adaptive control in Slotine and Li (1987) is adopted forthe inner-loop position control.

According to the adaptation procedure in Section 3, three steps are included in a single simulation.As discussed in Section 2.3, the exploration noise which meets the persistent excitation condition isrequired to guarantee the estimated parameters to convergeto the true values. It is well known thatthe sum of sinusoid waves is a typical swept frequency signalwhich meets the persistent excitationcondition. Therefore, in the simulation, the exploration noise is chosen asν =

∑10w=1

0.4w

sin(wt) andadded in the first step. The impedance model for this step isf = f0 + ν, wheref0 = −x − x + xd inSection 4.2 andf0 = −x + xd in Section 4.3. Impedance adaptation is conducted in the second step,and it stops when the condition‖Pk‖ < 10−10 satisfies. The initialPk isP0 = 10Ip, whereIp representsthep dimensional unit matrix. The impedance function for this step isf0. The desired impedance modelbased on the proposed method is obtained at the end of the second step, and it is used in the third step.

Corresponding to Sections 3.1 and 3.2, the following two environments are considered in two sub-sections:0.001x + 0.01x + (x − 0.2) = −f and0.01x + (x − 0.2) = −f . Note that0.2 is the initialposition of the robot arm. According to (22) and (30), ifA andB are known, the optimal solutions of theRiccati equations can be obtained and thus the desired impedance model (24) and (32) can be obtained.It is referred as “LQR”, and compared with the the proposed method in this paper. It is necessary to em-phasize thatA, B and thus the desired impedance model are only available in the simulation studies forthe comparison purpose, and they are usually unknown or needto be estimated in a typical application.This is the motivation of this paper, as already discussed inthe Introduction.

4.2 Mass-Damping-Stiffness Environment

In this subsection, the mass-damping-stiffness environment 0.001x + 0.01x + (x − 0.2) = −f isconsidered. In the first case, the weights in (15) are given byQ1 = 1, Q2 = 1 andR = 1. The desiredimpedance model (24) isf = −0.99x − 0.41x + 0.04xd based on knownA andB. The simulationresults are shown in Figs. 3, 4, 5 and 6. In Fig. 3, the positionof the robot arm in the Cartesian space isshown. In the first two seconds, there is a large position error between LQR and the proposed method.This is because there is exploration noise and the initial impedance modelf = −x−x+xd+ν is not thedesired model. The second step takes a very short time and thedesired impedance function convergesvery quickly. More details about the convergence performance can be found in Fig. 5, where the error ofimpedance parameters with respect to iteration number is shown. This error is defined as‖Kk−K‖, andit decreases to around 0.05 when the adaptation stops at the 5th iteration (each iteration takes 0.1s). Thedesired impedance model is obtained asf = −0.98x−0.38x+0.06xd with the proposed method, and itis used until the end of the simulation. It is found that the obtained impedance model is very near to butnot exactly the same as the desired one under LQR, i.e.,f = −0.99x− 0.41x+ 0.04xd. As a result, theposition in Fig. 3 is near to the position under LQR but there is still a small error. This is caused by theadaptation process in the inner-loop, and it is illustratedby Fig. 6 where the control performance of the

Page 12: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 11

inner-loop is shown. Particularly, although zero trackingerror in the inner-loop is achieved after about2s, the tracking error exists at the beginning of the simulation. Therefore, the dynamics of the robot armare actually not exactly governed by the given impedance model.

0 2 4 6 8 10 12 14 16 18 200

0.05

0.1

0.15

0.2

0.25

0.3

0.35

time (s)

posi

tion

(m)

xd

proposedLQR

Step 3: f=−Kkξ

Step 2: f=−K0ξ

Step 1: f=−K0ξ+ν

Figure 3.: Desired trajectory and actual trajectory,Q1 = 1, Q2 = 1 andR = 1

To further investigate the effect of inner-loop performance on the impedance adaptation, an additionalsimulation study is conducted. Instead of starting from thebeginning of the simulation, the impedanceadaptation starts at 2s when the inner-loop tracking error is near to zero. Simulation results are shownin Figs. 7 and 8. The desired impedance model is obtained asf = −0.98x − 0.40x + 0.04xd with theproposed method, which shows that the perfect adaptation can be achieved when a perfect tracking isguaranteed throughout the parameter adaptation period. Bycomparing the simulation results in Figs. 3,4 and Figs. 7, 8, we conclude that a good inner-loop performance during the parameter adaptation periodis crucial to guarantee a perfect impedance adaptation. When the good control performance of the inner-loop is not achieved during the parameter adaptation period, the adapted parameters cannot converge tothe exact values. Conversely, when a perfect inner-loop control performance is guaranteed during theadaptation process, the perfect impedance adaptation can be realized.

To further verify the effectiveness of the proposed impedance adaptation, another cost function ischosen in the second case. Particularly, the weights in (15)are given byQ1 = 1, Q2 = 10 andR = 1.Compared to that in the first case, the weight of the velocity is larger so it is expected that the systemresponse is slower. Similarly, the desired impedance model(24) is obtained asf = −3.15x − 0.41x +0.02xd based on knownA andB. The simulation results in this case are given in Figs. 9 and 10, andthe impedance model obtained with the proposed method isf = −3.14x − 0.40x + 0.02xd. While theposition in Fig. 3 converges to around 0.25m in 6s, the position in Fig. 9 converges to around 0.22m in10s. Similarly, the interaction force in Fig. 4 converges toaround -0.05N in 8s, and in Fig. 10 it convergesto around -0.02N in 16s. The above results are coherent with the expectation and verify the validity ofthe propose method. Similarly as in LQR, differentQ1, Q2 andR can be chosen to realize differentinteraction performances, e.g., either “softer” interaction or more accurate trajectory tracking.

Page 13: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

12 Shuzhi Sam Ge, Yanan Li and Chen Wang

0 2 4 6 8 10 12 14 16 18 20−0.05

0

0.05

0.1

0.15

0.2

time (s)

inte

ract

ion

forc

e (N

)

proposedLQR

Step 1: f=−K0ξ+ν

Step 2: f=−K0ξ

Step 3: f=−Kkξ

Figure 4.: Interaction force,Q1 = 1, Q2 = 1 andR = 1

1 2 3 4 5 6 7 8 9 100

0.1

0.2

0.3

0.4

0.5

0.6

0.7

iteration

erro

r of

impe

danc

e pa

ram

eter

s

Figure 5.: Error of impedance parameters,Q1 = 1, Q2 = 1 andR = 1

4.3 Damping-Stiffness Environment

The damping-stiffness environment0.01x + (x − 0.2) = −f is considered in this subsection. Thesimulation conditions are the same as in the previous subsection, which are described in Section 4.1. Theweights in (25) are given byQ2 = 1 andR = 1 and the desired impedance model (32) is obtained asf =−0.41x+ 0.70xd based on knownA′ andB′. Similar results as in the previous subsection are obtained.In particular, the impedance model obtained with the proposed method isf = −0.40x + 0.71xd. Theoptimal interaction performance is achieved as shown in Figs. 11 and 12.

Page 14: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 13

0 2 4 6 8 10 12 14 16 18 20−4

−2

0

2

desi

red

traj

ecto

ry(r

ad)

0 2 4 6 8 10 12 14 16 18 20−0.1

−0.05

0

0.05

0.1tr

acki

ng e

rror

(rad

)

0 2 4 6 8 10 12 14 16 18 20−0.1

0

0.1

0.2

0.3

time(s)

adap

tatio

n pa

ram

eter

s

Figure 6.: Inner-loop control performance

0 2 4 6 8 10 12 14 16 18 200

0.05

0.1

0.15

0.2

0.25

0.3

0.35

time (s)

posi

tion

(m)

xd

proposedLQR

Step 1: f=−K0ξ+ν

Step 2: f=−K0ξ

Step 3: f=−Kkξ

Figure 7.: Desired trajectory and actual trajectory,Q1 = 1, Q2 = 1 andR = 1

4.4 Discussion

In summary, it has been shown that the proposed method can be adopted to obtain a desired impedancemodel based on a given cost function which determines an optimal interaction performance. Comparedto impedance learning developed in the literature such as Kim et al. (2010), Buchli et al. (2011), the pro-posed method does not require the repetitive learning process and thus provides certain convenience. Theoptimal impedance adaptation can be achieved if the “perfect” tracking can be guaranteed in the inner-loop. As discussed in Remark 9, the proposed method in Jiang and Jiang (2012) may not be applicableto the scenario where the environment is changing rapidly. In the future work, we will investigate howto derive an impedance adaptation in face of dynamically changing environments. Besides, some issues

Page 15: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

14 Shuzhi Sam Ge, Yanan Li and Chen Wang

0 2 4 6 8 10 12 14 16 18 20−0.05

0

0.05

0.1

0.15

0.2

time (s)

inte

ract

ion

forc

e (N

)

proposedLQR

Step 1: f=−K0ξ+ν

Step 2: f=−K0ξ

Step 3: f=−Kkξ

Figure 8.: Interaction force,Q1 = 1, Q2 = 1 andR = 1

0 2 4 6 8 10 12 14 16 18 200

0.05

0.1

0.15

0.2

0.25

0.3

0.35

time (s)

posi

tion

(m)

xd

proposedLQR

Step 1: f=−K0ξ+ν

Step 2: f=−K0ξ

Step 3: f=−Kkξ

Figure 9.: Desired trajectory and actual trajectory,Q1 = 1, Q2 = 10 andR = 1

in real-world implementations might not be well described in the simulation studies, so future work willalso be dedicated to incorporate the proposed impedance adaptation with a real robot arm and furtherinvestigate the issues in real-world implementations (e.g., the effect of exploration noise, environmentmodel uncertainties and time delay) of the proposed impedance adaptation. Furthermore, as discussed inBraun et al. (2012b), it is nontrivial to find a proper cost function in many situations, which will be alsoone of the future works that we will focus on.

Page 16: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

International Journal of Control 15

0 2 4 6 8 10 12 14 16 18 20−0.05

0

0.05

0.1

0.15

0.2

time (s)

inte

ract

ion

forc

e (N

)

proposedLQR

Step 1: f=−K0ξ+ν

Step 2: f=−K0ξ

Step 3: f=−Kkξ

Figure 10.: Interaction force,Q1 = 1, Q2 = 10 andR = 1

0 2 4 6 8 10 12 14 16 18 20

0.16

0.18

0.2

0.22

0.24

0.26

0.28

0.3

0.32

time (s)

posi

tion

(m)

xd

proposedLQR

Step 2: f=−K0’ξ’

Step 3: f=−Kk’ξ’

Step 1: f=−K0’ξ’+ν

Figure 11.: Desired trajectory and actual trajectory,Q2 = 1 andR = 1

5 Conclusion

In this paper, impedance adaptation has been developed to obtain the desired impedance parameterssuch that the optimal interaction is realized subject to unknown environments. The dynamics of theunknown environments have been investigated and the interaction requirement has been described byminimizing a certain cost function which includes trajectory tracking and force regulation. Adaptive op-

Page 17: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

16 REFERENCES

0 2 4 6 8 10 12 14 16 18 20−0.08

−0.06

−0.04

−0.02

0

0.02

0.04

0.06

time (s)

inte

ract

ion

forc

e (N

)

proposedLQR

Step 2: f=−K0’ξ’

Step 3: f=−Kk’ξ’

Step 1: f=−K0’ξ’+ν

Figure 12.: Interaction force,Q2 = 1 andR = 1

timal control for unknown linear system has been employed asthe fundamental of the proposed method.The validity of the proposed method has been verified throughsimulation studies.

References

Arimoto, S., Kawamura, S., and Miyazaki, F. (1984a), “Bettering operation of dynamic systems by learning: Anew control theory for servomechanism or mechatronics systems,”Proceedings of the IEEE Conference onDecision and Control, 23, 1064–1069.

Arimoto, S., Kawamura, S., and Miyazaki, F. (1984b), “Bettering operation of robots by learning,”Journal ofRobotic Systems, 1, 123–140.

Bertsekas, D.,Dynamic Programming and Optimal Control, Vol. 1, Athena Scientific Belmont, MA (1995).Braun, D.J., Howard, M., and Vijayakumar, S. (2012b), “Optimal variable stiffness control: formulation and appli-

cation to explosive movement tasks,”Autonomous Robots, 33, 237–253.Braun, D.J., Petit, F., Huber, F., Haddadin, S., Smagt, P., Albu-Schaffer, A., and Vijayakumar, S. (2012a), “Optimal

Torque and Stiffness Control in Compliantly Actuated Robots,” Proceedings of the IEEE/RSJ InternationalConference on Intelligent Robots and Systems, pp. 2801–2808.

Buchli, J., Stulp, F., Theodorou, E., and Schaal, S. (2011),“Learning variable impedance control,”InternationalJournal of Robotics Research, 30, 820–833.

Buerger, S.P., and Hogan, N. (2007), “Complementary stability and loop shaping for improved human-robot inter-action,” IEEE Transactions on Robotics, 23, 232–244.

Cheah, C.C., and Wang, D. (1998), “Learning impedance control for robotic manipulators,”IEEE Transactions onRobotics and Automation, 14, 452–465.

Cohen, M., and Flash, T. (June 1991), “Learning impedance parameters for robot control using an associativesearch network,”IEEE Transactions on Robotics and Automation, 7, 382–390.

Colbaugh, R., Seraji, H., and Glass, K. (1991), “Direct adaptive impedance control of manipulators,”Proceedingsof the IEEE Conference on Decision and Control, pp. 2410–2415.

Colgate, J.E., and Hogan, N. (1988), “Robust control of dynamically interacting systems,”International Journalof Control, 48, 65–88.

Corke, P.I. (1996), “A Robotics Toolbox for MATLAB,”IEEE Robotics and Automation Magazine, 3, 24–32.Craig, J.J., and Raibert, M.H. (1979), “A Systematic Methodof Hybrid Position/Force Control of a Manipulator,”

Computer Software and Applications Conference, IEEE Computer Society, pp. 446–451.De Roover, D., Bosgra, O., and Steinbuch, M. (2000), “Internal-model-based design of repetitive and iterative

Page 18: Impedance adaptation for optimal robot–environment interactionsro.sussex.ac.uk/id/eprint/72085/1/ImpedanceAdaptation20May13.pdf · impedance adaptation is developed for two general

December 14, 2017 21:23 International Journal of Control ImpedanceAdaptation20May13

REFERENCES 17

learning controllers for linear multivariable systems,”International Journal of Control, 73, 914–929.Ge, S.S., Lee, T.H., and Harris, C.J.,Adaptive Neural Network Control of Robotic Manipulators, London: World

Scientific (1998).Hogan, N. (1985), “Impedance control: an approach to manipulation-Part I: Theory; Part II: Implementation; Part

III: Applications,” Transaction ASME J. Dynamic Systems, Measurement and Control, 107, 1–24.Huang, L., Ge, S.S., and Lee, T.H. (2006), “Position/force control of uncertain constrained flexible joint robots,”

Mechatronics, 16, 111–120.Jiang, Y., and Jiang, Z.P. (2012), “Computational adaptiveoptimal control for continuous-time linear systems with

completely unknown dynamics,”Automatica, 48, 2699–2704.Johansson, R., and Spong, M.W. (1994), “Quadratic Optimization of Impedance Control,”Proceedings of IEEE

International Conference of Robotics and Automation, 1, 616–621.Kim, B., Park, J., Park, S., and Kang, S. (2010), “Impedance learning for robotic contact tasks using natural actor-

critic algorithm,” IEEE Transactions on Systems, Man and Cybernetics-Part B: Cybernetics, 40, 433–443.Kirk, D.E., Optimal Control Theory: An Introduction, Prentice-Hall Inc., Englewood Cliffs New Jersey (1970).Lewis, F., and Vrabie, D. (2009), “Reinforcement learning and adaptive dynamic programming for feedback con-

trol,” Circuits and Systems Magazine, IEEE, 9, 32–50.Li, Y., Ge, S.S., and Yang, C. (2012), “Learning Impedance Control for Physical Robot-Environment Interaction,”

International Journal of Control, 85, 182–193.Li, Z.J., Ge, S.S., and Ming, A. (2007), “Adaptive robust motion/force control of holonomi-constrained nonholo-

nomic mobile manipulators,”IEEE Transactions on Systems, Man and Cybernetics-Part B: Cybernetics, 37,607–616.

Lu, W.S., and Meng, Q.H. (1991), “Impedance control with adaptation for robotic manipulations,”IEEE Transac-tions on Robotics and Automation, 7, 408–415.

Matinfar, M., and Hashtrudi-Zaad, K. (2005), “Optimization-based Robot Compliance Control: Geometric andLinear Quadratic Approaches,”The International Journal of Robotics Research, 24, 645–656.

Rahman, M.M., Ikeura, R., and Mizutani, K. (2002), “Investigation of the impedance characteristic of human armfor development of robots to cooperate with humans,”JSME International Journal Series C, 45, 510–518.

Slotine, J.J.E., and Li, W. (1987), “On the adaptive controlof robotic manipulators,”The International Journal ofRobotics Research, 6.

Stanisic, R.Z., and Fernandez, A.V. (2012), “Adjusting theparameters of the mechanical impedance for velocity,impact and force control,”Robotica, 30, 583–597.

Tsuji, T., and Morasso, P.G. (1996), “Neural Network learning of robot arm impedance in operational space,”IEEETransactions on Systems, Man and Cybernetics-Part B: Cybernetics, 26, 290–298.

Tsumugiwa, T., Yokogawa, R., and Hara, K. (2001), “Variableimpedance control with regard to working processfor man-machine cooperation-work system,”Proceedings of IEEE/RSJ International Conference on Intelli-gent Robots and Systems, pp. 1564–1569.

Tsumugiwa, T., Yokogawa, R., and Hara, K. (2002), “Variableimpedance control based on estimation of humanarm stiffness for human-robot cooperative calligraphic task,” Proceedings of the IEEE International Confer-ence on Robotics and Automation, pp. 644–650.

Uemura, M., and Kawamura, S. (2009), “Resonance-based Motion Control Method for Multi-Joint Robot throughCombining Stiffness Adaptation and Iterative Learning Control,” Proceedings of IEEE International Confer-ence on Robotics and Automation, pp. 1543–1548.

Wang, D., and Cheah, C.C. (1998), “An iterative learning-control scheme for impedance control of robotic manip-ulators,”The International Journal of Robotics Research, 17, 1091–1099.

Wang, F., Zhang, H., and Liu, D. (2009), “Adaptive dynamic programming: An introduction,”ComputationalIntelligence Magazine, 4, 39–47.

Watkins, C. (1989), “Learning from delayed rewards,”Cambridge University, Cambridge, England, Doctoral the-sis.

Werbos, P.J. (1990b), “Consistency of HDP applied to a simple reinforcement learning problem,”Neural Networks,3, 179–189.

Werbos, P.J. (1990a), “A menu of designs for reinforcement learning over time,”Neural Networks for Control, pp.67–95.

Werbos, P.J.,Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches, Vol. 15 (1992).Werbos, P.J. (2009), “Intelligence in the brain: A theory ofhow it works and how to build it,”Neural Networks, 22,

200–212.Xu, J., Viswanathan, B., and Qu, Z. (2000), “Robust learningcontrol for robotic manipulators with an extension to

a class of non-linear systems,”International Journal of Control, 73, 858–870.Yang, B.H., and Asada, H. (1996), “Progressive learning andits application to robot impedance learning,”IEEE

Transactions on Neural Networks, pp. 941–952.