You are on page 1of 6

2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)

Congress Center Hamburg


Sept 28 - Oct 2, 2015. Hamburg, Germany

A Dual Quaternion Linear-Quadratic Optimal Controller for Trajectory


Tracking
M. M. Marinho, L. F. C. Figueredo, B. V. Adorno

Abstract— This work addresses the task-space design prob- specification. Such solution is generally not easily derived,
lem of a linear-quadratic optimal tracking controller for and in some cases the IK may even result in impossible
robotic manipulators using the unit dual quaternion formalism. or in an infinite number of solutions [1]. Moreover, as
The efficiency, compactness, and lack of singularity of the
representation render the unit dual quaternion a suitable modern manipulator robots are often equipped with low
framework for simultaneously describing the attitude and the level controllers at joint level, task-space techniques have
position of the end-effector. Motivated by the advantages of the advantage of defining the control law directly at the
this kinematic description, we propose a new task-space linear- end-effector, making direct use of the available low level
quadratic optimal tracking controller in order to find an controllers instead of redesigning them.
optimal trajectory for the end-effector, providing a tool to
balance more conveniently the end-effector error and its task- Closely intertwined with those fields, efficient representa-
space velocity. This is possible because the kinematic control tions for rigid body transformations have been a recurring
problem using the dual quaternion transformation invariant topic in several studies concerning robotics. Considering
error can be reduced to an affine time-varying system. The coupled and singularity-free representations, the use of dual
proposed optimal tracking controller allows the compensation quaternions over homogenous transformations is less com-
of trajectory induced disturbances, as well as other modeled
additive disturbances and known bias. Simulation results with putationally demanding, being an efficient and compact form
different design parameters provide a performance overview, for representing rigid transformations [3]. In addition, control
in comparison with standard kinematic controllers with and laws are defined directly over a vector field, eliminating the
without a feed-forward term, for tracking a desired reference. need to extract additional parameters or to design matrix-
based controllers (i.e., controllers based on the matrix struc-
I. I NTRODUCTION ture of SE(3)). In the light of those advantages, there has
The study of modeling and control strategies for robotic been an increasing interest in works related to the design of
manipulators is closely related to the specificities of the kinematic controllers in dual quaternion space. Those works
desired task. Several tasks can be accomplished by stiff comprise rigid motion stabilization and tracking [4], robust
robotic manipulators at relatively low velocities and accelera- control [5], multiple rigid body coordination [6], multiple
tions, justifying the use of kinematic control strategies. Sim- arm manipulation [7], and human-robot interaction [8].
plifications that arise from the kinematic description allow Nonetheless, results in dual quaternion literature aim
designing controllers that do not require inertial parameter solely at stabilizing the trajectory of the rigid body, without
specifications of the manipulator. Applications of kinematic much concern on control effort and optimality. For instance,
controllers branch from the stable proportional gain feedback although stable the proportional gain controller [9] shows an
to several sub-fields of study as, for instance, robustness intrinsic delay in the presence of time-varying trajectories.
to singularities and perturbations, workspace optimization, Adding a feedforward term may eliminate the delay, but has
simultaneous execution of multiple tasks, and several forms the disadvantage of requiring a control effort that inevitably
of optimization [1]. In the present study, we take interest at grows alongside the trajectory time derivative. In this context,
optimization-based kinematic controllers at task-space. optimal criteria allow the designer to choose a performance
In practical scenarios, there exists a discrepancy between in an intermediate ground, with no phase shift and with
the task-space, where robot tasks are specified, and joint- optimal control effort.
space, where the actuation takes place. One of the main Among optimal and robust control of robot manipulators,
drawbacks of joint-space control techniques lies in this dis- earlier works consider mostly the dynamic model of the
crepancy [2], as they require an external inverse kinematics robot, using the computed torque control scheme. For in-
(IK) solution. Given a task specification at the end-effector, stance, [10] uses minimum variance control to find optimal
the IK provides a set of joint configurations that match that controllers considering input constraints, with results applied
to a two degree-of-freedom (DOF) planar manipulator. A
This work has been supported by the Brazilian agencies CAPES, CNPq, related work uses fuzzy neural networks to approximately
and FAPEMIG. compute the optimal control [11]. Another approach uses
M. M. Marinho is with University of Tokyo, Mitsuishi Sugita Lab-
oratory, Tokyo, Japan. Email: murilo@nml.t.u-tokyo.ac.jp. a single network adaptive critic controller, resulting in a
L. F. C. Figueredo is with Universidade de Brasília, LARA, Brazil. near optimal solution, which is tested using two DOF of
Email:figueredo@lara.unb.br. B. V. Adorno is with the Graduate a simulated WAM robotic manipulator [12]. The aforemen-
Program in Electrical Engineering - Federal University of Minas Gerais
(UFMG) - Av. Antônio Carlos 6627, 31270-901, Belo Horizonte-MG, tioned fuzzy controllers are able to show approximate results
Brazil. Email: adorno@ufmg.br. even without much knowledge of the manipulator dynamics.

978-1-4799-9994-1/15/$31.00 ©2015 IEEE 4047


They are motivated mostly because the dynamic parameters nion LQR can be used as a trajectory tracking controller,
are hard to obtain precisely, and vary with different pay- while also considering other modeled error disturbances. To
loads. Despite that advantage, the fuzzy controllers have demonstrate the implementation simplicity and efficiency, the
a degree of arbitrariness in parameter selection. Moreover, controller is evaluated using a simulated robotic manipulator
works using the computed torque technique optimize in with six DOF in a trajectory tracking task.
the manipulator joint space, requiring an external tool to
obtain the optimal IK. The implementation complexity also II. P RELIMINARIES
reduces most examples to two DOF robots. In a different We begin by recalling dual quaternions and their basic
approach [13], optimal results in the sense of convergence algebra when representing rigid transformations to establish
time are presented. In this context, instead of optimizing the notation used in this work, followed by a brief review of
the end-effector trajectory, the authors are interested in existing work in the kinematic control of robotic manipula-
bringing the end-effector to a final location in minimum tors.
time. Finally, a linear-quadratic optimization for manipulator
control has been found [14] considering the dynamic model A. Dual Quaternions
of a manipulator. The results mostly concern an efficient and Dual quaternions are the building blocks of the kinematic
practical trajectory smoothing algorithm, using the computed control theory implemented in this work. We begin with the
torque technique. However, most recent manipulator robots definition of the three imaginary components ı̂, ̂ and k̂ of
can be readily controlled by joint velocity inputs and less so a quaternion such that ı̂2 = ̂2 = k̂ 2 = ı̂̂k̂ = −1. Hence, a
through joint torque inputs. Moreover, as in other cases, the general quaternion h is given by h = h1 + h2 ı̂ + h3 ̂ + h4 k̂,
optimization is done in joint space, requiring an external IK and its conjugate is defined as h∗ , h1√ − h2 ı̂ − h3 ̂ − h4 k̂.
planner. The norm of a quaternion h is k h k= hh∗ .
An arbitrary rotation of a rigid body by an angle θ around
A. Contributions
an axis n = nx ı̂ + ny ̂ + nz k̂ is represented by the unit
In this work, we extend the results from dual quater- norm quaternion r = cos (θ/2) + n sin (θ/2). A translation
nion based controllers with the introduction of an optimal described by t = tx ı̂+ty ̂+tz k̂ can be associated to a rotation
quadratic tracking control law in the task space of the r in order to represent the complete rigid motion. This is
general serial-link manipulator. It compensates for modeled represented by the unit dual quaternion h = r + ε (1/2) tr,
perturbations of the pose error, while considering the control where ε is nilpotent; i.e., ε 6= 0 but ε2 = 0. Moreover, the
effort. The optimization-based controller is by itself a rele- conjugate of h is defined as h∗ , r ∗ + ε( 12 tr)∗ .
vant improvement on the literature as it provides the designer Since a general dual quaternion g is composed of eight
with a more intuitive set of performance indexes and extends elements, that is, g = g1 + g2 ı̂ + g3 ̂ + g4 k̂ + ε(g5 +
the range of applications of the dual quaternion formalism g6 ı̂ + g7 ̂ + g8 k̂), the vec operator is used to map it
to robotics. into an eight-dimensional column vector; i.e., vec g ,
Moreover, in contrast to standard optimal control tech-  T
niques for manipulators, the proposed criterion yields opti- g1 g2 g3 g4 g5 g6 g7 g8 . We also define the
mized task-space variables that greatly simplify the control matrix C8 , diag (1, −1, −1, −1, 1, −1, −1, −1) such that
implementation. As a consequence, instead of optimizing the vec h∗ = C8 vec h. Finally, given dual quaternions g 1 ,
− +
joint-space velocity, we aim at obtaining an optimal result g 2 , the Hamilton operators H (·), H (·) are transformation
for the end-effector velocity at task-space. The analysis is matrices satisfying the following relation [17]:
particularly relevant in advanced manipulation tasks—for   +  −  
instance, manipulation in hazardous or unstructured environ- vec g 1 g 2 = H g 1 vec g 2 = H g 2 vec g 1 . (1)
ments, and multi-arm manipulation—that require precise and
safe interaction in contrast to convergence speed. In the view B. Kinematic Control
of human-robot interaction, the speed of the end-effector The forward kinematics model (FKM) of a serial manipu-
largely influences the human acceptance of the contact [15]. lator robot (that is, the mapping between the n-dimensional
Indeed, even if the robotic system is able to prevent undesired vector of joint positions θ ∈ Rn and the end-effector pose
injuries, a human is likely to be in a state of constant stress h) can be obtained directly in dual quaternion space using
and discomfort [16]. A controlled speed at task-space also algebraic manipulations [17]. In addition, the differential
prevents drifts and overshoots caused by large accelerations FKM (i.e., the mapping between the joint velocities θ̇ ∈ Rn
requirements at the low level controllers. and the generalized end-effector velocity vec ḣ ∈ R8 ) is
To derive the optimal control criterion, we show that the given by
dual quaternion kinematic control can be reduced to an affine vec ḣ = Jθ̇, (2)
time-varying system with respect to the dual quaternion
transformation invariant error. This allows us to obtain a where J ∈ R8×n is the manipulator Jacobian (which is also
recursive expression for the optimal control using a linear- found algebraically [17]) and depends on the current robot
quadratic regulator (LQR). By modeling a continuous trajec- posture. As (2) is a linear mapping, it is common practice to
tory as the perturbation of the error, the proposed dual quater- design closed-loop controllers based on the pseudo-inversion

4048
of J that exponentially reduce the error between the current Consider that the manipulator has to track the trajectory
pose h and a constant desired pose hd ; that is, during t ∈ [0, tf ]. Therefore, we seek a feasible trajectory to
minimize the following cost function
θ̇ = KJ† vec(hd − h), (3)
1 tf T
Z
1 T
e Qe + uT Ru dt, (6)

where K is a positive definite matrix and J† is the singular F = e(tf ) Se(tf ) +
2 2 t0
value decomposition pseudoinverse of J [9].
given the matrices S, Q(t) ≥ 0 and R(t) > 0 with S, Q, R
III. D UAL Q UATERNION K INEMATIC LQR ∈ R8×8 . The matrix S is the weight of the final error norm,
whereas the time-varying matrices Q and R weigh the error
In this section we describe the novel optimal controller
cost along the trajectory and the control effort in terms of
which is the main contribution of this work. We begin with
end-effector velocity norm, respectively. As long as N is
the transformation invariant error definition together with
well conditioned, an increase in R will, ceteris paribus, also
the kinematic equations reviewed in the last section. We
cause an overall decrease in joint velocities. The constrained
show that the kinematic control with a time-varying reference
optimization of (6) leads to an optimal feedback without
from the point of view of the error reduces to a linear
excessive expenditure of control energy while keeping the
time-varying system with an additive perturbation term. By
error e(t) near zero [18].
using a suitable solution, we find the optimal state-feedback
For a continuously differentiable optimal trajectory
controller and its computation is discussed.
xopt (t) = (eopt (t), uopt (t)) along the constraint surface
First, consider an arbitrary continuous desired trajectory
hd , hd (t) in dual quaternion space and the current end- G(t) = A(t)e(t) + u(t) + c(t) − ė(t) = 0, (7)
effector pose h. The invariant error is given by [5]
the necessary and sufficient conditions for the global min-
e = 1 − h∗ hd . (4) imum are given by ∂F/∂u = 0 and ∂F/∂e = 0 as F is
convex. From the optimal trajectory, we also find that the
Since the LQR optimization computation assumes future
gradient ∇F of F is perpendicular to the constrained surface
knowledge of the explicit variables, we need to remove any
(7), that is, there must be no component of ∇F parallel to G
direct dependency on h. Hence,
at an optimal point—otherwise, it would be possible to move
e = 1 − h∗ hd =⇒ eh∗d = h∗d − h∗ =⇒ h∗ = h∗d − eh∗d . away from xopt and down the gradient of F while remaining
on (7), contrary to the assumption that xopt is a minimum
With time-varying h and hd , the derivative of (4) is given [19], [20]. In this sense, to constrain the optimization to a

by ė = −ḣ hd − h∗ ḣd . Therefore feasible trajectory, we define the augmented cost function
Z tf
∗ ∗
ė = −ḣ hd − (h∗d − eh∗d )ḣd = −ḣ hd + eh∗d ḣd − h∗d ḣd . H=F+ pT (Ae + u + c − ė)dt, (8)
t0

Using the vec operator on both sides and defining e , vec e where p is the Lagrange multiplier for the state equations.
yields Since ∇G defines the normal vectors to each component of
− −   G, the constrained trajectory is satisfied if ∇F +pT ∇G = 0,
ė = −H (hd ) C8 vec ḣ + H h∗d ḣd e − vec h∗d ḣd . and p defines the magnitude difference between ∇F and the
−   components of ∇G.
By defining A , H h∗d ḣd and c , −vec h∗d ḣd , using According to [20], [21], we have ∂H/∂u = 0 and
− ∂H/∂e = 0 as necessary and sufficient conditions for the
(2), and defining N , H (hd ) C8 J, we obtain
optimal trajectory, hence
ė = Ae − Nθ̇ + c. ∂H
= 0 =⇒ u = −R−1 p. (9)
∂u
Instead of considering θ̇ as the input signal for the system,
we can consider as input the end-effector velocity u using In addition, we use the Leibniz integral rule in (8), that is,
the mapping u = −Nθ̇. This allows the optimization to be Z tf Z tf
done in task-space variables. We focus on finding the optimal − pT ėdt = pT (t0 )e(t0 ) − pT (tf )e(tf ) + ṗT edt,
controller for the affine time-varying system t0 t0

to find ∂H/∂e = 0, which in turn implies


ė(t) = A(t)e(t) + u(t) + c(t). (5)
ṗ = − Qe + AT p .

(10)
Therefore, since the error dynamics is an affine system, we
can use a LQR to find the optimal solution by choosing a Please note that ∂ 2 H/∂u2 must be positive to minimize
proper cost function. Notice that the time-varying trajectory (8), which requires R > 0. The system and proposed cost
disturbance in the error dynamics is given by c(t). Other function allow the use of the costate function [21]
modeled continuous disturbances can also be grouped into
c(t) and used in the same solution. p(t) = Pe + ξ, (11)

4049
where P is a time-varying proportional gain and ξ is a cost. Since the system input is the end-effector velocities, a
weighted feedforward term. The derivative of (11) yields proper choice of the elements of R allows the designer to
˙ define end-effector velocities into which is more costly to
ṗ(t) = Ṗe + Pė + ξ. (12)
move over others.
After using (11) in (9) and substituting the result in (5), and An interesting characteristic of the LQR design is the
using (11) in (10), we replace the resulting equations in (12) ability to define time-varying weights in the optimization
to find process. For instance, we may devise a trajectory that begins
far from the end-effector initial position. To avoid large joint
− ξ˙ − AT ξ + PR−1 ξ − Pc velocities when the error is expected to be large, we may
assign a smaller Q at the beginning of the trajectory.
 
= Ṗ + PA + AT P − PR−1 P + Q e. (13)
IV. S IMULATIONS AND D ISCUSSION
Considering that (13) must hold for any choice of initial state
e and that both P and ξ do not depend on the initial error, In order to demonstrate the controller behavior under
we need simultaneously different sets of parameters, a simulated task was devised.
( The controller is implemented using the DQ_robotics1 in
Ṗ + PA + AT P − PR−1 P + Q = 0 MATLAB and the simulated robot is the COMAU Smart
Six with six DOF, with which we wish to perform a complex
ξ˙ + AT ξ − PR−1 ξ + Pc = 0,
motion to demonstrate the effectiveness and simplicity of the
which means controller.
( We begin by choosing a periodic translation in all axes,
Ṗ = −PA − AT P + PR−1 P − Q
(14)
ξ˙ = −AT ξ + PR−1 ξ − Pc. td (t) = (0.99707 + 0.1 cos(πt)) ı̂
The boundary conditions to solve (14) are obtained using + (0.1 + 0.1 cos(πt)) ̂ + (1.075 + 0.1 sin(πt)) k̂.
the final time, tf , of our trajectory. We set the feed-forward
In order to generate a motion that modifies all rotational
term to zero in the terminal time—that is, ξ(tf ) = 0—to
degrees of freedom, we also choose a varying end-effector
find the first boundary condition. From (11) with ξ(tf ) = 0,
orientation given by
we find that ∂H/∂e(tf ) = 0 yields P(tf ) = S.
− 
r d (t) = cos (φ/2) + sin (φ/2) v k v k−1 ,

Note that we know A(t) = H h∗d ḣd for all t, as it
depends only on the desired trajectory. We can then numer- with φ = 0.25 sin(πt) + π and v = cos(πt)ı̂ + sin(πt)̂ +
ically solve the differential Ricatti equation P(t) backwards (sin(πt) − 2)k̂. The trajectory is the combined motion given
in time. As c = −h∗d ḣd is also known for all t and, with by
the solution of P(t), we can find ξ(t) by also solving it hd (t) = r d (t) + ε 12 td (t)r d (t).
backwards in time.
Therefore, from (9) and (11) the optimal control is given  The robot initial configuration
T is given by θ0 =
by u(t) = −R−1 (Pe + ξ) . Applying as joint velocities, we 0 0 −π/2 0 π/2 0 , with the end-effector placed
find 40 cm away from the desired trajectory.
θ̇ = N† R−1 (Pe + ξ) . (15) Three LQR controllers are designed so that their perfor-
mance can be evaluated. In order to simplify the choice of
A. DQ-LQR Computation and Parameter Selection parameters, we define s, q, r ∈ R, such that
The system (14) containing the Ricatti equation is nonlin-
S = sI, Q = qI, R = rI,
ear, and the closed-form solution may not be found for all
cases. As a consequence, finite differences approximations where I ∈ R8×8 is an identity matrix. The choice of
are normally used; that is, parameters is closely intertwined with the task specificities.
d P(t + τ ) − P(t) For instance, we suppose that the end-time error is not of
P(t) ≈ higher importance than the error in the remainder of the
dt τ
trajectory by setting s = 0. By knowing that the end-effector
for a small sampling time τ that directly affects the approx- starts far from the desired trajectory, we design a trajectory
imation precision. error weight q that begins small and increases over time:
In order to obtain P(t) for all t, we begin with P(tf ) and q = 0.001 for t ≤ 0.5, linearly increasing to q = 1 for
use the derivative approximation backwards in time using the t ∈ (0.5, 1.0], and q = 1 otherwise. All three controllers
known values of A(t), which depends only on the desired share the same s and q.
trajectory, up to P(0). The same can be applied to find an In order to compare performance, we distinguish the three
approximate solution for ξ(t), by also considering c(t). All controllers by choosing a decreasing weight for the control
these calculations can be done before the system starts to effort parameter, that is r ∈ {0.1, 0.01, 0.001}. The overall
operate. behavior of the three optimal controllers can be seen in
The design parameters of the controller are the matrices
Q, S and R. Matrices Q and S are related to the state norm 1 http://dqrobotics.sourceforge.net

4050
Fig. 1. In all controllers, the choice of q allows a well- the end-effector pose error. The controller allows the designer
behaved initial motion. Moreover, there is no noticeable to set optimal gains for a given trajectory in terms of control
phase shift after 1.0 s, when q reaches its final value. The effort and error regulation, and may be applied in both non-
effect in the decreasing of r is also clearly shown in Fig. 1, redundant and redundant manipulators. We also provided
that is, a higher control effort yields smaller trajectory error. simulation results using a six DOF manipulator to illustrate
We also compare the performance of the proposed optimal the performance and effectiveness of the proposed optimal
controllers with classic control methods. To this aim, we also criterion in comparison with a proportional gain controller
simulated a proportional controller (K controller) and a proportional gain controller with feed-forward.
θ̇ = N† Ke
and a proportional controller with a feed-forward term (K +
FF), given by R EFERENCES
† ∗ [1] M. W. Spong, S. Hutchingson, and M. Vidyasagar, Robot Modeling
θ̇ = N (Ke − vec h ḣd ),
and Control. New York: John Wiley & Sons, 2005.
[2] O. Khatib, “Manipulator control at kinematic singularities: a dynami-
where K = kI with k ∈ R+ for both controllers. cally consistent strategy,” in Proceedings IEEE/RSJ International Con-
The performance of the optimal controller with r = 0.001, ference on Intelligent Robots and Systems. Human Robot Interaction
in comparison with the K and K+FF controllers with k = 10, and Cooperative Robots, vol. 3. IEEE Comput. Soc. Press, 1995, pp.
84–88.
is shown in Fig. 2. It is noticeable that whereas the K [3] M. Schilling, “Universally manipulable body models—dual quaternion
controller shows the expected phase shift in spite of the high representations in layered and dynamic MMCs,” Autonomous Robots,
control effort, the proposed criterion with r = 0.001 closely vol. 30, no. 4, pp. 399–425, 2011.
[4] X. Wang and C. Yu, “Unit dual quaternion-based feedback lineariza-
resembles the K + FF controller, but without the initial peak tion tracking problem for attitude and position dynamics,” Systems &
velocity. This allows the LQR controller to send smoother Control Letters, vol. 62, no. 3, pp. 225–233, Mar. 2013.
joint velocities when compared to the abrupt velocities seen [5] L. Figueredo, B. Adorno, J. Ishihara, and G. Borges, “Robust kine-
matic control of manipulator robots using dual quaternion repre-
in the K and K + FF controllers. It also yields smaller sentation,” in Proceedings of the IEEE International Conference on
accelerations at the end-effector which, among other benefits, Robotics and Automation. Karlsruhe: Germany, 2013, pp. 1949–1955.
prevents damages to the manipulator and may reduce the [6] X. Wang, C. Yu, and Z. Lin, “A Dual Quaternion Solution to
Attitude and Position Control for Rigid-Body Coordination,” IEEE
discomfort during human-robot interaction [16]. Transactions on Robotics, vol. 28, no. 5, pp. 1162–1170, Oct. 2012.
The overall results show that by choosing the design [7] L. Figueredo, B. Adorno, J. Ishihara, and G. Borges, “Switching
parameters accordingly, we are able to define a trade-off strategy for flexible task execution using the cooperative dual task-
space framework,” in IEEE/RSJ International Conference on Intelli-
between control effort and tracking error along the trajectory. gent Robots and Systems, 2014, pp. 1703–1709.
For further information, the integral norm of the end-effector [8] B. V. Adorno, A. P. L. Bó, and P. Fraisse, “Kinematic modeling and
velocity, acceleration, and tracking error for each controller control for human-robot cooperation considering different interaction
roles,” Robotica, vol. 33, no. 2, pp. 314–331, Feb. 2015.
is shown in Table I. From Fig. 2, it is clear that the smaller [9] H.-L. Pham, V. Perdereau, B. Adorno, and P. Fraisse, “Position
error from the K+FF controller is a consequence of smaller and orientation control of robot manipulators using dual quaternion
convergence time, which in turn is obtained in exchange of feedback,” in Proceedings of the IEEE/RSJ International Conference
on Intelligent Robots and Systems. Taipei: Taiwan, 2010, pp. 658–
abrupt velocities and accelerations—respectively, 22.5% and 663.
260% larger than the results from the LQR with r = 0.001. [10] M. Negm, “Application of optimal preview and adaptive controllers for
robotics manipulator with control input constraints,” in Proceedings of
TABLE I the IEEE International Symposium on Industrial Electronics, vol. 2,
1999, pp. 853–860 vol.2.
I NTEGRAL NORM OF THE END - EFFECTOR VELOCITY (V EL .),
[11] R.-J. Wai, C.-H. Tu, and K.-Y. Hsieh, “Design of intelligent optimal
ACCELERATION (ACC .), AND TRAJECTORY ERROR (E RR .). tracking control for robot manipulator,” in 2003 IEEE/ASME Interna-
tional Conference on Advanced Intelligent Mechatronics, vol. 1, July
LQR K K+FF 2003, pp. 478–483 vol.1.
r=0.1 r=0.05 r=0.001 k=10 k=10 [12] S. Dutta and L. Behera, “Policy iteration based near-optimal control
Err. 1.134 0.886 0.450 0.582 0.176 scheme for robotic manipulator with model uncertainties,” in IEEE
Vel. 3.297 3.833 4.857 5.820 5.954 International Symposium on Intelligent Control, Aug 2013, pp. 358–
Acc. 0.028 0.033 0.048 0.176 0.173 363.
[13] M. Galicki and D. Ucinski, “Time-optimal motions of robotic manip-
ulators,” Robotica, vol. 18, no. 6, pp. 659–667, 2000.
[14] O. Egeland and E. Lunde, “Trajectory generation for manipulators
V. C ONCLUSIONS using linear quadratic optimal tracking,” in IEEE International Con-
ference on Robotics and Automation, Apr 1988, pp. 376–381.
In this paper, we studied the design of a dual quaternion [15] A. Mertens, C. Brandl, I. Blotenberg, M. Ludtke, T. Jacobs, C. Brohl,
kinematic finite-time linear-quadratic optimal tracking con- M. P. Mayer, and C. M. Schlick, “Human-robot interaction: Test-
troller. By deriving a perturbed time-varying linear system ing distances that humans will accept between themselves and a
robot approaching at different speeds,” in Ambient Assisted Living,
from the transformation invariant error definition, we pro- ser. Advanced Technologies and Societal Change, R. Wichert and
posed an optimal criterion strategy to control the error in H. Klausing, Eds. Springer Berlin Heidelberg, 2014, pp. 269–286.
relation to perturbations caused by a time-varying trajectory. [16] P. A. Lasota, G. F. Rossano, and J. A. Shah, “Toward safe close-
proximity human-robot interaction with standard industrial robots,” in
The resulting optimal technique can be readily applied to IEEE International Conference on Automation Science and Engineer-
account for other modeled sources of bias and disturbance to ing, 2014, pp. 339–344.

4051
Reference Control effort
Quaternion angle
r=0.001

Angle [rad]

Generalized velocity norm


2
r=0.05
3.2
r=0.1
3 1.5

1
End effector translation Quaternion axis
0.4
x-axis [m]

1 0.2 0.5

x-axis
0
0.9 -0.2
-0.4 0
0.8 0 1 2 3
Simulated time [s]
End effector trajectory
0.2
y-axis [m]

y-axis
0.1 0.5
0
0 1.15

z-axis [m]
1.1
z-axis [m]

-0.2
z-axis
1.1 -0.4 1.05
-0.6
-0.8 0
1 1
0.1
0 1 2 3 0 1 2 3
0.8 0.9 1 0.2 y-axis [m]
Simulated time [s] Simulated time [s] x-axis [m]

Fig. 1. Linear-quadratic optimal controller with different parameters; the curves show the end-effector translation (left) and orientation (center), the
control effort (top-right), and end-effector trajectory (bottom-right). The trade-off between error and control effort is noticeable, and the end-effector motion
shows no steady-state phase shift in relation to the reference.

Reference Control effort


Quaternion angle
LQR, r=0.001
Angle [rad]

Generalized velocity norm


K, k=10 3.3
3.2
K + FF, k=10 3.1 10
3
2.9

End effector translation Quaternion axis


5
0.4
x-axis [m]

1 0.2
x-axis

0
0.9 -0.2
-0.4 0
0.8 0 1 2 3
Simulated time [s]
End effector trajectory
0.2
y-axis [m]

y-axis

0.1 0.5
0
0 1.15
z-axis [m]

1.1
z-axis [m]

-0.2 1.05
z-axis

1.1 -0.4
-0.6
-0.8 0
1 1
0.1
0 1 2 3 0 1 2 3
0.8 0.9 1 0.2 y-axis [m]
Simulated time [s] Simulated time [s] x-axis [m]

Fig. 2. Comparison between the optimal controller and the proportional gain controllers; the curves show the end-effector translation (left) and orientation
(center), the control effort (top-right), and end-effector trajectory (bottom-right). The initial peak velocity using the proportional controllers is noticeable,
which results from the large initial error. Without the prediction term, the proportional controller has an inherent phase shift in relation to the desired
trajectory.

[17] B. V. Adorno, “Two-arm manipulation: from manipulators to enhanced Journal, vol. 6, no. 11, pp. 2586–2591, Nov 1968.
human-robot collaboration [Contribution à la manipulation à deux
bras : des manipulateurs à la collaboration homme-robot],” Ph.D.
dissertation, Université Montpellier 2, 2011.
[18] A. Locatelli, Optimal Control: An Introduction. Birkhäuser Basel,
2001.
[19] J. Wyatt, “Lagrange multipliers, constrained maximization, and the
maximum entropy principle,” University Lecture Notes, June 2004.
[20] R. M. Murray, “Optimization-based control,” in Feedback Systems,
K. J. Åström and R. M. Murray, Eds. Princeton Univ. Press, 2010.
[21] V. Garber, “Optimum intercept laws for accelerating targets.” AIAA

4052

You might also like