Impact Factor 3.392

The #1 largest and #1 most cited open-access publisher of neuroscience

This article is part of the Research Topic Movement in 3D space


Front. Behav. Neurosci., 20 February 2013 |

3D kinematics using dual quaternions: theory and applications in neuroscience

  • 1Institute of Information and Communication Technologies, Electronics and Applied Mathematics, Université Catholique de Louvain, Louvain-la-Neuve, Belgium
  • 2Institute of Neuroscience, Université Catholique de Louvain, Brussels, Belgium
  • 3Centre for Neuroscience Studies, Queen's University, Kingston, ON, Canada
  • 4Canadian Action and Perception Network, Toronto, ON, Canada

In behavioral neuroscience, many experiments are developed in 1 or 2 spatial dimensions, but when scientists tackle problems in 3-dimensions (3D), they often face problems or new challenges. Results obtained for lower dimensions are not always extendable in 3D. In motor planning of eye, gaze or arm movements, or sensorimotor transformation problems, the 3D kinematics of external (stimuli) or internal (body parts) must often be considered: how to describe the 3D position and orientation of these objects and link them together? We describe how dual quaternions provide a convenient way to describe the 3D kinematics for position only (point transformation) or for combined position and orientation (through line transformation), easily modeling rotations, translations or screw motions or combinations of these. We also derive expressions for the velocities of points and lines as well as the transformation velocities. Then, we apply these tools to a motor planning task for manual tracking and to the modeling of forward and inverse kinematics of a seven-dof three-link arm to show the interest of dual quaternions as a tool to build models for these kinds of applications.

1. Introduction

Our environment is 3-dimensional (3D) and our body can be represented as a rigid multibody system evolving in that 3D environment, with different body parts moving relative to each other. Accurately describing the 3D kinematics of such systems is important for diverse applications in neuroscience which involve 3D kinematics: (1) forward kinematics of different 3D systems: the position and/or orientation of an end-effector (a tool held by the hand for example), (2) 3D Reference frame transformation or reference frame encoding [for example: 3D vestibulor-ocular reflex (VOR), 3D visuomotor transformation, 3D eye-head gaze shifts, 3D spatial updating, …], and (3) inverse kinematics: the set of joint angles corresponding to a given end-effector position and velocity. In practice, these models allow us to analyze 3D data acquired through behavioral experiments as well as to make predictions.

To model the kinematics of a rigid multibody system, we use the framework of the geometric algebra introduced by David Hestenes (see the textbook Hestenes, 1986), which easily applies to and is very convenient to model mechanical systems, in particular the kinematics. Hestenes applied this approach to some eye (Hestenes, 1994a) and arm reaching (Hestenes, 1994b) movements. Actually, he chose the geometric algebra of quaternions which have been sometimes used to model 3D eye, head and arm movements (see for example: (Tweed and Vilis, 1987; Straumann et al., 1991; Hore et al., 1992; Haslwanter, 1995; Crawford and Guitton, 1997). Quaternions allow us to easily deal with rotations but not with translations. Yet we have to deal with translations since the origins of reference frames attached to each body generally do not coincide. For example, the eye rotation center is offset from the head rotation center. In order to deal with translations, in addition to the rotation, we will use the geometric algebra of dual quaternions (see Bayro-Corrochano, 2003). Dual quaternions also allow to elegantly represent a screw motion (Funda and Paul, 1990b; Aspragathos and Dimitros, 1998), defined as a combined rotation and translation along the rotation axis. They have been used in Blohm and Crawford (2007) to model combined 3D eye and head rotations in the context of 3D visuomotor transformation for reaching movements. While dual quaternions have not been widely used in the neuroscience community, they have been applied in other fields. Indeed, they have been employed with success in computer vision (Walker et al., 1991; Phong et al., 1993; Goddard and Abidi, 1998; Torsello et al., 2011) or in robotics (Daniilidis, 1999; Bayro-Corrochano et al., 2000).

The goal of this paper is to provide a tutorial of the dual quaternion geometric algebra to the neuroscientists and then to show its interests and advantages to several applications in sensorimotor control. First, we summarize the theory necessary to introduce the different applications. The different dual quaternion operations are described and we also provide our MATLAB implementation of these operations in supplementary materials for the potential interested reader. Then we describe several applications using the dual quaternion formalism.

There are several reasons to use dual quaternions to represent the 3D kinematics. First, it is a compact and geometrically meaningful way of representing 3D rotations, translations and screw motions (Funda and Paul, 1990a; Kim and Kumar, 1990; Aspragathos and Dimitros, 1998; Daniilidis, 1999; Kavan et al., 2008). Furthermore, there are no singularities when representing 3D rotations (this is not the case if a Fick, Helmoltz or Euler representation is chosen), because a rotation is represented through a single angle θ and the axis of rotation n (Chou, 1992; Grassia, 1998). Moreover computational requirements (in terms of storage and speed) are lower than for other methods, including homogeneous coordinates (Funda et al., 1990; Funda and Paul, 1990b; Walker et al., 1991; Aspragathos and Dimitros, 1998), which could be important for behavioral experiments needing real-time data and processing (for example, a forward kinematic model applied to on-line raw data) but also for post-processing treatments in data analysis and model predictions. In an implementation perspective again, the neuroscience researcher concerned with the practical implementation of the transformations using dual quaternions will most likely want to consider many transformations simultaneously. For example, when analyzing experimental data, or when making predictions (from a few thousand to many more). In this context, dual quaternions are much more convenient than homogenous matrices for example. Indeed, we easily deal with an arbitrary number N of simultaneous dual quaternions transformations (using standard matrix operations), while for homogeneous coordinates, it requires the use of 3D tensors, which is possible but not easy to implement. Yet neuroscience researchers want a tool efficient and easy to use at the same time. Last but not least the dual quaternion formalism can be used to model point and lines transformations (Bayro-Corrochano, 2003), which is not the case for other formalisms like homogeneous coordinates (only point transformations). Yet applications sometimes demand that we combine both types of transformations (see section 3.3 of this article) and dual quaternions provide a unified and easy way to do that. We would have had to employ two different approaches to perform kinematic operations on points and lines if we had not used dual quaternions.

2. Theory

In this section, we provide a tutorial or short description of the dual quaternion algebra based on the literature (Hestenes, 1994a; Bayro-Corrochano, 2003). The reader familiar with these concepts can skip this part and move to section 3. However, as far as we know of, many readers in the sensorimotor science community are not familiar with dual quaternions and therefore they will find in this section many useful dual quaternions operations which can be useful to model their 3D problem in neuroscience. Moreover, we provide a MATLAB implementation for each of these operations, for the reader who would like to use dual quaternions but who perhaps would be restrained by the implementation. This toolbox is available for download at the followings urls:

In the following, we introduce the dual quaternion algebra, based on the quaternion algebra. Then we introduce the rotation and translation operations using the dual quaternion algebra. Afterwards we describe how points and lines can be encoded using dual quaternions, before explaining how applying the rotation or translation on these objects. Then we describe what a screw motion is and how it is encoded using a dual quaternion. Finally we briefly talk about the implementation and conversion to other formalisms.

2.1. Basics

2.1.1. Quaternion geometric algebra

First let us define the geometric product of two 3D vectors a and b (see Hestenes, 1994a; Bayro-Corrochano, 2003). It is denoted ab and can be expressed as:


where a · b is the classical dot product yielding a scalar number and ab is a new entity called a bivector, resulting from the wedge product of a and b. This wedge product is antisymmetric: ab=ba. We will see later how to compute it in practice. Now let us consider three orthonormal basis vectors in a right-handed reference frame, σ1, σ2, and σ3. From the definition of the geometric product and the antisymmetric property of the wedge product, we can get that:

σiσj={ 1ifi=jσiσjifij

Therefore, we obtain three basis bivectors: σ1σ2, σ2σ3, and σ3σ1. Then the unit right-handed pseudoscalar i is defined (see Hestenes, 1994a) as:


i is thus an entity which is called a trivector. The i symbol is used by analogy with the complex numbers. Indeed, using the properties of the geometric product and the definition of i, we have: i2 = −1. i is a duality operator since we can notice that:


Therefore, the three basis bivectors are dual entities to the three basis vectors and vice-versa, which implies that for every bivector a, there exists a corresponding vector a such that a=ia. In the following, we will always use the bold notation for bivectors, and top arrows for vectors. In particular, the bivector obtained by the wedge product can be expressed as:


where a×b is a vector representing the classical cross product of the vectors a and b. A quaternion is the sum of a scalar and a bivector (Hestenes, 1994a):


and can be obtained as the geometric product of two vectors, as shown by Equation (1). The sum of two quaternions A=a0+ia and B=b0+ib is a quaternion C=c0+ic computed as follows:


The multiplication of two quaternions A=a0+ia and B=b0+ib is a quaternion C=c0+ic. It is computed as (see details in Appendix A):

C=AB=(a0+ia)(b0+ib)=(a0b0a·b)c0      +i(a0b+b0aa×b)c

Other quaternion operations are also useful. The conjugate of a quaternion A is defined by:


The norm of a quaternion A is denoted: ||A||=(a02+a · a). The inverse of a quaternion A is denoted A−1 and defined as A1=A*||A||2.

2.1.2. Dual quaternion geometric algebra

A dual quaternion D is defined by


where D0 and D1 are two quaternions and ϵ is a mathematical operator with the property that:


The multiplication of two dual quaternions D = D0 + ϵD1 and E = E0 + ϵE1 is another dual quaternion F = F0 + ϵF1 and using the property (2), F is equal to:


We notice that the dual quaternion multiplication involves three quaternion multiplications.

Several conjugates exist for the dual quaternion D = D0 + ϵD1, which are used depending on the operations needed. The first one is obtained by applying the quaternion conjugate to each quaternion composing the dual quaternion.


The conjugate of a dual number can also be used:


And both operations can also be combined:


The following identities are also useful:

(AB)*=B*A* (AB¯)=A¯B¯(AB¯)*=B¯*A¯*

The norm of a dual quaternion D is a dual number (a dual number is of the form a = a0 + ϵa1 where a0 and a1 are real scalar numbers and ϵ is the same operator as for dual quaternions, with the property that ϵ2 = 0). The dual quaternion norm is computed in the following way (see Kavan et al., 2008):

||D||=||D0||+ϵ[ D0*D1+D1*D02||D0|| ]scalar

where ||D0|| is the quaternion norm of D0 and [Q]scalar is the scalar part q0 of a quaternion Q = q0 + q. The inverse of a dual quaternion A = A0 + ϵA1 can be computed and is written (see Kavan et al., 2008):


If the non-dual part, A0, of the dual quaternion A has a zero norm, then A has no inverse.

In the next sections, the dual quaternions representing rotations, translations and or screw motions, as well as points and lines, are described using unitary dual quaternions, i.e., dual quaternions with a norm equal to 1. They are unitary under the condition that:


Therefore, unit dual quaternions belong to a 6-dimensional manifold and are specified by six different parameters. From now on, we always use the bivector notation in order to avoid writing the i symbol at each equation. Anyway, remember that each bivector a is directly related to its counterpart vector a by the relation a=ia.

2.2. Rotations and Translations

In this section, we describe the dual quaternions associated with rotations and translations, as well as their velocity: rotational velocity, translational velocity. These operators will be used and combined later to carry out complex kinematic transformations.

2.2.1. Rotation

By Euler's theorem, any 3D rotation may be described by a unique unitary rotation axis n and a single rotation angle θ. In the dual quaternion framework, a pure rotation (whose rotation axis goes through the origin of the reference frame) is described by:


where n is a bivector representing the unitary rotation axis coordinates expressed in the reference frame and θ is the rotation angle. Actually, Rrot is a quaternion, or a dual quaternion with a dual part (the part premultiplied by ϵ) equal to 0.

2.2.2. Translation

A pure translation is defined by its unitary axis t and its distance d. A translation dual quaternion Ttrans is written under the following form:


2.2.3. Rotational velocity

Potential users should pay attention to the fact that in 3-D, the rotational velocity quaternion is not the derivative of the rotation quaternion (this is also the case if classical rotation matrices are considered). It has been shown by Hestenes (1986) that the derivative of the rotation quaternion with respect to time can be written as:


where Ωrot=Ωrot=iΩrot is a rotational velocity dual quaternion which has a dual part equal to zero, and which has a non-dual part with a zero scalar component. Ωrot represents the rotational velocity. Kinematically, the norm of Ωrot, ||Ωrot|| represents the instantaneous angular velocity while the normalized vector Ωrot||Ωrot|| the instantaneous rotation axis. Also, it can be shown that:


2.2.4. Translational velocity

The translational velocity dual quaternion is derived from the translation dual quaternion Ttrans. Differentiating the expression of the translation dual quaternion Ttrans, we obtain:

T˙trans=0+ϵ(d˙(t)2t(t)+d(t)2t˙(t))          =ϵv2

where v=˙d˙(t)t(t)+d(t)t˙(t).

2.3. Description of Point and Line Positions and Velocities

Here we describe how we can describe either points (location and velocity) or lines (location and velocity) using the dual quaternion formalism.

2.3.1. Point position and velocity

Following what is done by Bayro-Corrochano (2003) for example, a 3D physical point P with coordinates x in a given reference frame R is represented by the dual quaternion X = 1 + ϵx.

The velocity of point P is represented by the dual quaternion V = = ϵ.

2.3.2. Line position and velocity

Dual quaternions provide a convenient way to represent lines (Daniilidis, 1999; Bayro-Corrochano, 2003). A line can be represented by six coordinates (Plücker coordinates), specifying the line orientation n and the position of an arbitrary point of the line in space, p. The line representation is moreover independent of the choice of the point p. A line dual quaternion L is written:


We can notice that m is indeed independent of the choice of p since the cross product p×n depends only on the component of p orthogonal to n, and this orthogonal component is obviously the same for every point on the line L. Note also that a line dual quaternion is unitary.

The line velocity, L˙, is obtained by computing the derivative of the line dual quaternion L:

L˙=n˙+ϵm˙  =n˙+ϵd(pn)dt  =n˙+ϵ(p˙n+pn˙)(6)

We see that the line velocity dual quaternion is related to the rate of change of n, as well as the rate of change of p. Note that Equation (6) does not depend on the arbitrary choice of p and its rate of change (see proof in Appendix B).

2.4. Kinematic Operations on Points and Lines

Now we describe the kinematic transformations on point and lines, as the dual quaternion formalism allows us to use the same kinematic operators for point or line transformations, which is a major advantage.

Let us consider the reference frame R′, with the same origin as reference frame R, but lying in a different orientation. The transformation between both reference frames is a pure rotation with axis n and angle θ. Then, a point P has the following coordinates in the R′ frame (passive rotation):


where Rrot is the rotation quaternion. Rrot¯* is the conjugate quaternion: Rrot¯*=cos(θ2)sin(θ2). For rotations dual quaternions, Rrot¯*=Rrot*, since rotation dual quaternions have a dual component equal to zero. Therefore, in the following, we use R*rot instead of Rrot¯* since it is shorter to write.

From another point of view, we may want to know the new position of a mobile point P after a rotation Rrot applied to a reference position P0 = 1 + ϵx0 in a given reference frame R (an active rotation):


Let us now assume that reference frame R″ has the same orientation as R but R″ origin is offset from R origin. The transformation between both reference frames is a simple translation along a given unitary axis t and distance d. The unitary axis t is parallel to the line connecting the origins of frames R″ and R and is directed from R″ origin toward R origin. In R″ frame, point P has coordinates:


where Ttrans is the translation dual quaternion. The conjugate dual quaternion that we use for point transformation (see Bayro-Corrochano, 2003) is Ttrans¯*=1ϵ(d2t)=Ttrans. We can also apply the same reasoning for the active translation of a given point.

The velocity of the 3D physical point P with coordinates x in a given reference frame R is represented by the dual quaternion V=˙X˙=0+ϵx˙. It can be of interest to compute the velocity of P in the frame R′, which is itself rotating compared to R. In that reference frame R′, the velocity dual quaternion V′ is:

V=d(RrotXRrot*)dt   =Rrot.XRrot*+RrotXRrot*˙+RrotX˙Rrot*(10)

Using relationship (4), expression (10) may be developed:


We can observe that if frame R′ does not move compared to frame R, Ωrot = 0, then the only difference is that V′ has its coordinates expressed in the R′ frame instead of the R frame.

From another point of view, we may want to know the velocity of a mobile point P during a rotation Rrot(t) applied to a reference position P0 = 1 + ϵx0 with reference velocity 0 in a given reference frame R (an active rotation):

P˙=d(Rrot*P0Rrot)dt  =12(PΩrotΩrotP)+Rrot*P˙0Rrot

Let us now compute the velocity V″ of point P in the reference frame R″ that is translated with respect to R. The transformation between both reference frames is a simple translation motion along a given unitary axis t and distance d(t). In R″ frame, the velocity of point P is:

V=d(TtransXTtrans)dt    =Ttrans.XTtrans+TtransXTtrans.=ϵd˙t+TtransX˙Ttrans=X˙    =0+ϵ(x˙+d˙t)               (11)

For some applications, we may be interested in applying the kinematic operators to lines instead of points. For example, we may be interested in the orientation of an end-effector (in addition to the position) and/or the way this orientation changes with time. Lines provide a useful way to answer those questions. An example will be described in the applications. The operations are exactly identical to those described for point transformations except that another dual quaternion conjugate is used: the quaternion conjugate, A*, instead of the mixed conjugate, A*¯ (see Equations 7, 9). This slight difference is due to the way that points and lines are encoded with dual quaternions (Bayro-Corrochano, 2003).

2.5. Screw Motion

Before moving to the applications, we describe a last (less known) kinematic operator: a screw displacement, or screw motion. First we explain what a screw motion is and how this kinematic operation is encoded with dual quaternions. Then we derive the screw motion velocity. Finally we show how screw motion can be applied easily to both points and lines.

The reader should be aware that the screw motion is not a revolutionary kinematic operator, but it is a simple alternative, providing a single dual quaternion directly for an operation which can be also obtained by combining a translation along and a rotation around an axis line which is offset from the origin. Whatever the approach chosen by the user, the final result for your application will be the same. However, screw motion dual quaternions provide a more compact way to write the kinematics for this type of movement.

2.5.1. Screw motion definition

A screw motion describes a rotation of angle θ about a line whose direction is specified by a unitary axis n and whose location can be described by an arbitrarily chosen point on the line, with coordinates a, followed/combined by a translation of distance d along this axis n. a is an arbitrary point of the line and this line does not necessarily go through the origin, such that the screw motion formalism also can be used to describe rotations about eccentric points—in contrast to quaternions and rotation matrices, which can only characterize rotations about the origin. A screw motion can be represented by the following dual quaternion:


Note that the quantity an is independent of the vector a we choose on the line to describe the line location. Equation (12) is derived (see Daniilidis, 1999; Bayro-Corrochano, 2003) by starting from the rotation quaternion R describing a rotation of angle θ around the unitary axis n. However, as previously noted, the quaternion formulation implicitly assumes that the rotation axis line passes through the reference frame origin. If the rotation line passes through a point with coordinates a, a general expression for the rotation dual quaternion whose axis is offset by a, RT, is obtained by applying a translation operator TL=1+ϵa2 to the quaternion R, by left-multiplying it by the conjugate of TL and right-multiply it by TL. It is quite similar to the translation operation that we apply to lines (instead of rotation dual quaternion here) and that we described in section 2.4 :

RT=TL*RTL    =(1ϵa2)(cosθ2+sinθ2)(1+ϵa2)    =(cosθ2+sinθ2)+ϵ((an)sinθ2)

The dual quaternion RT describes the same rotation but takes the offset between the reference frame origin and the line into account. RT is a transformation of R and therefore TL had to be applied from both sides. To obtain the screw motion dual quaternion M, we need to combine a translation T=1+ϵd2n along the line axis to the rotation RT (here we combine operators instead of transforming one operator, so there is only a simple left-multiplication):

M=TRT=(cosθ2+sinθ2)        +ϵ(d2sinθ2+nd2cosθ2+(an)sinθ2)

2.5.2. Screw motion velocity

The screw motion velocity dual quaternion, , is derived from the screw motion dual quaternion expression, M, in Equation (12). Indeed, a screw motion can be decomposed into a translational part and a rotational part, as shown in section 2.5.1. We have:



M˙=T˙TL*RTL+TTL*.RTL+TTL*RTL.+TTL*R˙TL    =T˙Rtranslationalterm+TL*.R+RTL.offsetterm+12TTL*RΩTLrotationalterm

where the second line is obtained through a simple calculation (translation has no effect on a velocity dual quaternion). We can observe the contributions of several terms related to the translational velocity along the screw axis, the translational velocity of the screw axis itself and the rotational velocity term.

Similarly to rotation and translation operations, we may easily apply screw motion operations on points or lines.

2.5.3. Screw motion applied to points

Let us consider now an active screw motion M = RTT applied to a reference point P0. The resulting position is then:

P=M¯*P0M  =TRT¯*P0RTT(14)

If we consider again the active screw motion M = RTT applied to the reference point P0 with velocity 0, the resulting velocity is then:

P˙=d(M¯*P0M)dt  =M¯*.P0M+M¯*P0M˙+M¯*P˙0M

As will be seen in the applications, we can linearly combine several rotations, translations and screw motions in a compact expression, facilitating the geometrical interpretation and the clarity.

2.5.4. Screw motion applied to lines

Let us describe the final parameters of the line L = n + ϵm resulting from applying a screw motion described by M = RTT to a reference line L0 = n0 + ϵm0. Compared to the expression used for points (see Equation 14), the only slight difference is the use of another conjugate, as previously mentioned (see Bayro-Corrochano, 2003).


Knowing the structure of a line dual quaternion (see Equation 5), we can extract the line parameters n and m. Since m=pn where p is any point located on line L, we can choose a particular p by adding an additional constraint. An example would be to take the point p which is the closest to the origin.

We can also compute the line velocity L˙=n˙+ϵm˙ resulting from a screw motion M to a reference line L0 with velocity L˙0.

L˙=d(M*L0M)dt  =M˙*L0M+M*L0M˙+M*L0.M

Again, we can retrieve the parameters and . Having retrieved a particular point p of the line, we can extract its velocity (orthogonally to the line orientation n).

In the Appendix C, we also describe some useful dual quaternions identities which may be used in the diverse sensorimotor applications that we will present in section 3. In the next section, we discuss the implementation of dual quaternions.

2.6. Implementation of Dual Quaternions

From an implementation perspective, we can write a quaternion A=a0+ia as a four-element vector Atab:


where a=(a1a2a3)T where T is the transpose matrix operator. If we consider N quaternions simultaneously, we just create a 4 × N matrix where each column is a the vector representation of the corresponding quaternion.

Using this representation, the quaternion multiplication is implemented using matricial operations:


We can also consider N simultaneous multiplications of two quaternions. This is achieved by adapting the above formula to matrices instead of vectors. These simultaneous operations are not just for theoretical purposes or generalizations. Indeed, again addressing to the neuroscientists communities, we often have to deal with hundreds or thousands of trials in behavioral experiments. Often, we compare our data to model predictions. In this framework, let us consider that we use quaternions for our model predictions. In this case, it is much faster to transform N quaternions (corresponding to N trials) simultaneously than to run them individually through a for loop.

A dual quaternion D=D0+ϵD1=d0+id0+ϵ(d1+id1) can be represented using a 8-dimensional vector Dtab (which can also be considered as the juxtaposition of two 4-dimensional vectors representing the two quaternion components of the dual quaternion):


Again, we can generalize to N dual quaternions by using a 8 × N matrix. The dual quaternion multiplication is implemented using matricial operations and the quaternion multiplication defined in Equation (15):


We will not describe how all dual quaternion operators and transformation are implemented, as most of them are easy or just a consequence of the encoding of a general dual quaternion that we just described. However, for the reader who would not like to implement herself/himself all the dual quaternion operations in her/his favorite language, we provide in supplementary materials a dual quaternion toolbox written in MATLAB providing functions implementing all the dual quaternions operations previously mentioned. Furthermore, an example and a read me file are also available. A document listing several quaternion and dual quaternion Matlab toolboxes developed by others is also provided. In this way, the potential user has access to our toolbox but also to others, and therefore he can judge which one is the most suitable for him/herself. We wanted to develop our own Matlab toolbox because the other ones did not gather all the functionalities we needed.

After introducing the dual quaternion algebra in this first part and how it can be applied in kinematics, we now move to the second part of this manuscript. This part describes how dual quaternions and their derivatives can be used to easily describe several applications from sensorimotor control in neuroscience.

3. Applications

The dual quaternion algebra is very convenient to express the motion of rigid bodies, especially our body parts. In the following, we describe several applications of this theory. First we describe the reference frame transformations required in the 3D visuomotor transformation for reaching and tracking movements. Then we describe the forward and inverse kinematics problem for a two-link arm focusing only on the end-effector position. For that problem, we will use point transformations. Finally we focus on the forward and inverse kinematics problem for a three-link arm holding a tool whose position and orientation matters. We will use both point and line transformations. One huge advantage of dual quaternions is the fact that we can use them for both point and line transformations, which is not the case with other formalisms (homogeneous matrices are not suitable for line transformation for example, while dual matrices were developed to tackle line transformations and not point transformations).

3.1. Reference Frame Transformation: Movement Planning

Here we study reference frame transformations in the context of visuomotor transformations. For instance, visually guided arm movements to reach for a seen object. Indeed, the brain has to transform the visual information about the target of interest into a set of motor commands for the arm muscles. To this end, the brain plans the movement ahead (see Shadmehr and Wise, 2005). However, the transformation between retinal information and the spatial motor plan is not trivial (see Blohm and Crawford, 2007), since our eyes and head move relative to the body (and thus relative to the shoulder, insertion point of the arm). For instance, let us consider the motion of a target in space. If the head is rolled toward the left or right shoulder, the projection of the spatial motion onto the retina will be different depending on the head roll angle (Leclercq et al., 2012). Therefore, the brain should take the head roll into account in order to generate an accurate motor plan for the arm. This transformation amounts to expressing the retinal motion into a spatial motion, thus it is a reference frame transformation problem.

In the following we describe two transformations. First we describe how we express the retinal position and motion of a target, e.g., a tennis ball, as a function of the spatial trajectory of this target and the 3D eye-head-shoulder geometry. This transformation is useful for a neuroscientist dealing with behavioral experiments. Indeed, the spatial target position and velocity are often specified by designing the experiment (by choice of the experimenter), while the retinal position and motion can not be measured directly. These signals need to be estimated using the transformation model that we develop in this first part and measured and/or known signals about the 3D eye-head-shoulder kinematics. Then we describe the inverse transformation which computes the spatial position and velocity of a target from the knowledge of the retinal position and velocity as well as the 3D eye-head-shoulder geometry. This transformation is important from a neuroscience perspective as it is the transformation that the brain should implement in order to generate a spatially accurate arm movement (see Blohm and Crawford, 2007; Leclercq et al., 2012). Furthermore, using this theoretical transformation, we can easily test hypotheses about the availability, accuracy or precision of a signal (retinal and/or extra-retinal) in the brain. For that last application, we use the retinal position and velocity estimated through the first transformation described above.

In this context, dual quaternions provide a useful tool to express these transformations since they provide a geometrically meaningful way of expressing them, and they are easily implemented using a dual quaternion toolbox (as we provide it in the supplementary materials).

3.1.1. Computing the retinal position and motion

Figure 1 depicts a situation where a subject is confronted with a moving object (e.g., a ball), the target (denoted P), and she/he is going to interact with it (track or catch the target). Here we describe how the retinal position and velocity of the target are expressed as a function of the spatial position and velocity, using the dual quaternion formalism to express the 3D kinematics. In an experiment, we typically specify P position and velocity in a spatial reference frame, for example, the environment reference frame, or the shoulder reference frame in this case since we assume the shoulder remains fixed in the environment. So we need to estimate what the position and the velocity of the projection of P onto the retina are.


Figure 1. The picture represents a front view of the eye-head-shoulder system. The pointing target position (black star) and velocity (dashed arrow) are represented. These position and velocity can be represented in different reference frames, represented on the figure. These reference frames are: SS (shoulder-centered shoulder-fixed), HH (head-centered head-fixed), and EE (eye-centered eye-fixed). The offsets between the rotation centers are also represented: the the head position in eye-centered head-fixed coordinates, and tsh the shoulder position in head-centered shoulder-fixed coordinates.

In order to compute the P projection trajectory onto the retina, it is obvious that we need to know P kinematics in shoulder (space) coordinates and also what the kinematics of the eye-head-shoulder rigid body system are. In addition to a spatial reference frame, a right-handed orthonormal reference frame is attached to each rigid body (the head, the eye and the shoulder in this example, see Figure 1). Let us assume that P trajectory is known in the shoulder reference frame, PSS(t) = (PSSx, PSSy,PSSz). PSS(t) refers to P position in the Shoulder-centered Shoulder-fixed reference frame (-centered refers to the reference frame origin while -fixed denotes that the orientation of the axes is constant with respect to the specified rigid body) as a function of time.

The first step is to compute P trajectory (position and velocity) in an eye-centered eye-fixed reference frame, denoted PEE(t) and PĖE. By combining eye-in-head rotation (REH quaternion), offset between eye and head rotation centers (THE dual quaternion), head-on-shoulder rotation (RHS quaternion), and offset between head and shoulder rotation centers (TSH dual quaternion), we obtain the following expression for P position in eye-centered eye-fixed coordinates as a function of P position in shoulder-centered shoulder-fixed coordinates (using point transformations, as described in section 2.4; see also Appendix E for the derivation).


where the rotation quaternions are expressed as a function of the rotation angles and unitary axis.

For the P velocity, we differentiate Equation (16). And after a few calculations (see Appendix E), we obtain:


where PEH (resp.PHS) is the target position expressed in eye-centered head-fixed (resp. head-centered shoulder-fixed) coordinates. They can be computed, similarly to the expression in Equation (16), as:


There are three terms in the right side of this expression: the first one depends on the eye-in-head rotational velocity, ΩEH. The second one depends on the head-on-shoulder rotational velocity, ΩHS. The third term is related to the target velocity in the shoulder-centered shoulder-fixed reference frame, PṠS. When the eye-head-shoulder configuration is static during a trial, only the last term remains. In this case, the only change in velocity is due to the different orientations of the reference frames.

The first step for computing the retinal position and motion was to express the target position and velocity in a reference frame centered on the eye and fixed to the eye, PEE(t) and PĖE. Then the second step is to compute the projection of the target position onto the retina and the resulting velocity of this projection, which is the retinal position and velocity available for further processing in the brain. Computationally, it amounts to project PEE = 1 + ϵ PEE onto a sphere of radius reye (see Figure 2A). For simplicity, we assume reye = 1 (it is just a scaling factor).




where dP PEE is the distance of the target (in eye-centered coordinates). The velocity of the projection is computed as follows:


The resulting dual quaternion is proj˙PEE=ϵproj˙PEE. Note that if the motion of the target is spherical with the eye as center (isodistance trajectory), then d˙P=0, and the expression simplifies to proj˙PEE=PE˙EdP. However, most of the time, the distance is not constant, for example, if the target is moving in a frontoparallel plane. d˙P can be expressed as (see Appendix F):


Figure 2. Geometry of the velocity projection (A) and velocity reconstruction (B). (A) Projection of P position and velocity onto the eye (actually we plot the reversed retinal projection and velocity projection, to be directly comparable to the spatial position and velocity). References to the reference frame are omitted on the figure for the sake of clarity. The two components which add to obtain proj˙PEE are represented : yes. (B) Reconstruction of P position and velocity. The two components which add to obtain EE are represented: yes.

In this first part, we have computed the retinal target projection vector in eye-centered eye-fixed coordinates from the knowledge of target motion in spatial coordinates. This transformation is useful because we do not know the projection vector in retinal coordinates in advance. Therefore, from an experimental or simulation point of view, this transformation is important. Moreover the above transformation could be useful for the brain if the target is also known to the brain in spatial (or also head) coordinates (for example: auditory target, proprioceptive target in addition to the visual information about the target). Indeed, multisensory integration is carried out using multiple reference frames (McGuire and Sabes, 2009) and therefore auditory (head coordinates) information should be expressed in eye coordinates. This transformation could also be useful if the brain was to implement a forward model, predicting the sensory consequences of the arm motor command, in retinal coordinates.

Now we move to the inverse transformation, which computes the target motion in spatial (or shoulder) coordinates, with the retinal position and velocity as input. This transformation must be carried out by the brain in order to specify a spatially accurate motor plan for the arm (see Blohm and Crawford, 2007; Leclercq et al., 2012).

3.1.2. Computing the spatial position and velocity

Now we describe the model with projPEE and proj˙PEE as inputs, and PSS and PṠS as outputs. These outputs would be used to specify the motor plan (direction, velocity) of an arm tracking movement for example (Leclercq et al., 2012). Using this model, predictions about the motor plan can therefore be easily computed under different hypotheses of incomplete transformation. These hypotheses are made by asking whether extra-retinal signals (3D eye and head positions, 3D eye and head velocities) are available to the brain and whether they are biased or not, and whether they are highly or not much variable (in the framework of bayesian estimation for example).

We start the complete transformation by reversing the above described model. From projPEE and proj˙PEE, we first compute PEE and PĖE (see Figure 2B). Theoretically,


We observe that dP is necessary to reconstruct PEE. It is interesting to consider how the brain estimates dP. Most of the time, our vision is binocular and dP can be estimated with the use of binocular cues (see Blohm et al., 2008). However, vision is sometimes monocular (e.g., Leclercq et al. (2012) patched one eye in order to have only monocular vision). In this case, no binocular clues are available to estimate target depth. But monocular cues and a priori information also help to have an estimation of the target depth, but they are quite reduced for point-like target movements in complete darkness (see Leclercq et al., 2012). A paradigm of tracking in depth would be interesting to test, in monocular and binocular situations. That would allow to differentiate between monocular and binocular clues in order to estimate depth in the context of arm movements.

Theoretically, we also have:


thus d˙P also has to be estimated for the estimation of the target velocity in space.

To compute the target position in spatial coordinates, we invert Equation (16), using the fact that the kinematic dual quaternions are unitary and therefore, for any unitary dual quaternion A, we have A−1 = A*. Applying that, we have:


where TEH=THE*=1ϵthe2 and THS=TSH*=1ϵtsh2

For the target velocity in spatial coordinates, we inverse the relationship (17) and obtain:


These two transformations are implemented by the brain for the planning of spatially accurate movements, as it has been shown by Blohm and Crawford (2007) and Leclercq et al. (2012). These transformations are useful for the experimenter, in order to estimate the retinal position and motion of a target, and also to make different predictions on the motor plan generated by the brain, depending on hypotheses made about the signals available to the brain (accuracy, precision). However, we believe that dual quaternions are just a tool to easily model this transformation, and they are most likely not used by the brain as a way to implement these transformations. Most likely, these transformations are implemented in a distributed way by a neural network. Blohm et al. (2009) and Blohm (2012) show how these transformations could be achieved by the brain in a distributed way using a biologically inspired artificial neural network. Moreover they relate the properties of the artificial neurons to those of real neurons. Let us mention that dual quaternions are used in these theoretical studies to generate a huge number of input–output pairs (retinal motion as input and spatial motion as output).

3.2. Two-Link Arm Movements: Forward and Inverse Kinematics for End-Effector Position

We will now consider another example of active movements: 3D arm movements. In section 3.2, we focus on the end-effector trajectory of a two-link arm, describing its forward kinematics and the inverse kinematics using the dual quaternion formalism applied to point transformations. In section 3.3, we consider a three-link arm holding a tool (a screwdriver for example) to emphasize the end-effector orientation importance in several everyday life situations. The dual quaternion formalism is again used in order to model the forward and inverse kinematics for the orientation and position of the end-effector, but it is also applied to lines, in order to deal with the orientation transformation.

3.2.1. Forward kinematics for end-effector position

In the following, we consider the forward kinematics of the end-effector position of a two-link arm (see Figure 3). The forward kinematics consists in computing the end-effector position (and sometimes orientation) from the knowledge of the joints kinematics (rotation angles and axes, joint velocities). The advantage of dual quaternions to represent such transformations is the compactness and geometrical significance to express the joint kinematics.


Figure 3. Two-link arm top view. The shoulder insertion point, elbow and end-effector location are represented. The different reference frames are also represented: body-fixed shoulder-centered (BS, in red), upper arm -fixed shoulder-centered (US, in blue), and the lower arm -fixed elbow centered (LE, in green). The rotation angles of the shoulder and elbow are respectively denoted φ and θ.

Assuming a length of a (resp. b) for the upper arm (resp. lower arm), the position PBS of the end-effector P (B for body-fixed coordinates and S for shoulder-centered coordinates) can be expressed as:


where S stands for shoulder, U for upper arm, E for elbow and L for lower arm. RUB represents the upper arm rotation around the shoulder compared to a reference position (rotation angle: φ = 0). RLU represents the lower arm rotation around the elbow compared to a reference position (θ = 0). TES=1+ϵ12(a00)T represents translation dual quaternion associated with the offset between the elbow and the shoulder, in shoulder-centered upper-arm fixed coordinates. PLE is the (fixed) position of point P in lower-arm fixed elbow-centered coordinates:


Note that we can represent the forward kinematics of the end-effector if we know the parameters of the shoulder and elbow rotations (for example, angle and rotation axis), whatever the degrees of freedom.

We can also express the velocity BS of the end-effector P:


where we obviously assume that LE = 0. ΩUB and ΩLU represent the rotational velocities of the upper arm around the shoulder and of the lower arm around the elbow.

3.2.2. Inverse kinematics from end-effector position

Estimating the joint kinematic parameters from the knowledge of the end-effector position and velocity (and sometimes the orientation, see next) is a much more difficult task in general. This process is called inverse kinematics. It can be difficult for several reasons. First, we could solve the inverse kinematics equations (e.g., Equation 19) for the joint kinematic parameters, but theses equations are highly non-linear and no general method exists to solve this problem. For a particular case, it is sometimes possible and advantageous to compute analytical solutions, but it can not always be done. Numerical methods are also widely used to solve this problem, and in the following we will use this approach, which allows us to develop one general method for several different problems. Then, the kinematic system is usually redundant (like the human arm), meaning that there is an infinity of joint configurations that can achieve the prescribed end-effector kinematics. Finally, there are some geometrical configurations in which there are singularities and which tend to complicate the resolution of the inverse kinematics problem. This is a well studied problem in the field of robotics (Klein and Huang, 1983; Cheng and Gupta, 1991; Wang and Chen, 1991; Sciavicco and Siciliano, 1996; Tolani, 2000). Here, we take the two-link arm as an example and develop a methodology from our dual quaternion formalism to compute the inverse kinematics numerically.

The inverse kinematics problem is complicated, especially because the degree of freedom (dof) exceeds the dimension of the end-effector motion (this is called redundancy). For the two-link arm, we consider four degrees of freedom (three at the shoulder and one at the elbow), while the end-effector motion is only 3D. Therefore, there is an infinity of solutions to the problem.

First, we simplify the problem: we assume that the two-link arm moves in the horizontal plane, which reduces the joint dof to 2, yielding a well-posed problem for inverse kinematics. In this case, the velocity dual quaternion representing the 2D velocity of point P may be written as (see details in Appendix G.1):


where n is the rotation axis of the shoulder and the elbow (since we only consider planar motions) and where JDQ is an array composed of two point velocity dual quaternions (from an implementation perspective, JDQ is an 8 × 2 array, see section 2.6). We can build the 2 × 2 jacobian matrix J by extracting the 2D vector part of the velocity dual quaternion composing each column of JDQ (see Equation 20):

J=([ 12(PBSnnPBS) ]2Dvec[ RUB*PUEnnPUE2RUB ]2Dvec)

The dual quaternion equation (20) can then be expressed in a matrix form:


where BS is the 2D velocity vector component of the velocity dual quaternion BS. We then compute φ˙ and θ˙ as a function of BS to solve the inverse kinematics problem, assuming we know the current position of the joints, θ and φ, since the jacobian matrix J depends on these parameters: J = J(θ, φ). Indeed, J depends on PBS, and PBS depends on the rotation dual quaternions RUB and RLU (see Equation 17). Since θ is the rotation angle of RLU and φ is the rotation angle of RUB, it shows that J depends on θ and φ. Also, RUB and PUE depend on φ and appear in the expression of J.

However, we can solve for φ˙ and θ˙ by inverting J only if J is invertible. It is the case in most geometrical configurations of the two-link arm (indeed, J depends on φ and θ, so does the rank of J), except when θ = 0 (or θ = 180°), which represents a situation where the upper and lower arms are aligned. In this case, J is not invertible (its rank is equal to 1), and we can not apply the above method. This situation can correspond to two cases. If BS has a component parallel to the aligned arm, then there are no solutions for φ˙ and θ˙. If BS is orthogonal to the aligned arm, then there exists an infinity of solutions (in which case we can express the general form of the solution using the pseudo-inverse formalism, see later in the manuscript).

We notice that the differential velocity relationship linking the joint velocities and end-effector velocity is linear if we assume that θ and φ are known, and that will be the case in the numerical approach we use to solve the inverse kinematics problem. In order to solve the inverse kinematics problem for the joint positions, we can numerically integrate the joint velocities across time, knowing the joint positions at time t0 (an initial condition is needed). The simplest numerical integration is the Euler method:


where Δt is the integration step and should not be too large to achieve reasonable accuracy. Then, the new rotation dual quaternions can be updated as a function of θ(t + Δt) and φ(t + Δt) and we can iterate this process.

Now we want to consider movements in 3D space, not restricted to the plane. In this context, the shoulder has three-dof and the elbow has one-dof. It is clear that there are too many dof compared to the 3D motion of the end-effector, and intuitively many shoulder-elbow configurations will lead to the same end-effector position and velocity. In the following we explain one approach which can be taken to obtain one specific solution for the joints, as well as a general formula to express the set of all possible solutions.

The idea is to start with the dual quaternion expression of the velocity of the end-effector (e.g., Equation 19) and rewrite it in a matrix expression. We move from the dual quaternion representation to the matrix notation because we use tools from the matrix algebra (e.g., the pseudo-inverse matrix) to derive the inverse kinematics results (see below).

The relationship between the end-effector velocity BS and the rotational velocities of the shoulder, ΩUB, and the elbow, ΩLU, is described by Equation (19). Using the fact that: 12(PBSΩUBΩUBPBS)=ϵ(PBSΩUB) (see Appendix G.2) we can show that Equation (19) can be expressed as a linear matrix expression (see Appendix G.3 for the derivation):


where BS is the dual bivector part of BS, RMUB is the rotation matrix associated with the rotation dual quaternion RUB [see Equation (D.2) in Appendix D for how to compute this rotation matrix] and A˜PBS is the anti-symmetric matrix of rank 2 associated with the cross product: if v=a×b, it can also be written v=A˜ab where:


where a = (a1 a2 a3). Note that this matrix has always rank 2 since for an equation in x: a×x=v, the component of x parallel to a can be arbitrary. All the solutions x lie on a line.

Coming back to Equation (22), we see that the matrix J is of size 3 × 6 and we want to solve for the unknowns ΩUB and ΩLU. Since we consider one-dof at the elbow, the vector ΩLU must be aligned with a specific rotation axis nelbow = [0 0 1] (in upper arm fixed coordinates) since we assume that the elbow rotates around an axis which is orthogonal to both the lower and upper arms (see also Figure 3 to see why this axis is along the z-axis). This constraint can be written as nelbow×ΩLU=0. Therefore, we add this constraint to the kinematics equation (22) to obtain:

(A˜PBS  RUBMA˜PUE0  A˜nelbow)J(ΩUBΩLU)=(P.BS0)        (23)

Now this modified matrix J has generally rank 5 since A˜nelbow has rank 2 (but the rank can be even lower in certain geometrical configurations of the two-link arm, as explained previously with the two-dof two-links arm). Indeed, because we can rotate the two-link arm system around an axis linking the shoulder and the end-effector, without changing the end-effector position and velocity, there is an infinity of solutions (ΩUB, ΩLU) which lie in a one-dimensional manifold, which is mathematically defined as the kernel of J, i.e., the set Ker(J) = {x ∈ ℝ6 such that Jx = 0}. Therefore, since there is an infinity of solutions, we need some criterion that may be optimized to choose one optimal solution (according to this criterion), while Equation (23) is a constraint for the optimization problem.

The matrix J−1 is not defined and thus we choose to use a generalized inverse matrix (which exists for any matrix): the pseudo-inverse [see Klein and Huang (1983) for the use of pseudo-inverse in inverse kinematics], also called the Moore–Penrose inverse. We will denote it by J+. Therefore, one solution of the redundant system given by Equation (23) is given by:


This solution is actually the solution with minimal norm (Klein and Huang, 1983; Sciavicco and Siciliano, 1996), which minimizes the joint velocities in the context of the four-dof two-link arm. The general solution can also be expressed (see Sciavicco and Siciliano, 1996):


where w can be any vector in ℝ6, and I is the identity matrix of order 6 (for this example). In practice, Equation (24) requires that we explicitly compute the pseudo-inverse J+. The interested reader can refer to Appendix G.4 to see how this can be done.

Once we have a solution for the joint velocities x=(ΩUBΩLU), we can numerically integrate the joint velocities across time, knowing the joints positions at time t0. Several methods for numerical integration exist (see for example, Atkinson, 1989) and are used in the context of inverse kinematics (see Cheng and Gupta, 1991). Here, we will use the simple Euler method to update the rotation quaternions RUB and RLU. It is a bit more tricky than in the simple two-dof two-link arm (see Equation 21). Indeed, given that R˙=12RΩ, we can update the rotation dual quaternion as:


but the major problem is that in general R(t + Δt) is not a rotation quaternion anymore, and we have to normalize it by the norm of R(t + Δt) to ensure it is a rotation dual quaternion again. From a computational perspective, Funda et al. (1990) showed that the normalization operation is carried out much faster with quaternions than with rotation matrices, which is one of the advantages of using quaternions over rotation matrices. One alternative for the computation is the following. In angular vector notation, the magnitude of the vector represents the rotation angle while the normalized unit vector represents the rotation axis. Using this fact, the rotational displacement between time t and t + Δt is characterized by the angular vector ΩΔt which is of the form θn. Then, this angular vector can be expressed as a rotation dual quaternion (see Equation 3), ΔR. Then,


However, by using such a numerical integration scheme, errors arise since the numerical integration is not perfect. These errors propagate from one iteration to the other and the reconstructed end-effector location (using the forward kinematic model, see Equation 18) will drift from the real end-effector position (see Sciavicco and Siciliano, 1996). In order to avoid this problem, we take the position error between the reconstructed end-effector position, P^BS, and the real (or desired) end-effector position, PBS, into account. Actually, we apply the correction scheme described in Sciavicco and Siciliano (1996). The vector error at time t is:

e(t)=PBS(t)P^BS(t)      =PBS(t)FK(RLU(t),RUB(t))

where FK(.) is the forward kinematic function described by Equation (18). We take this error into account by adding a vector term proportional to the error in Equation (23):


where K is a positive definite (usually diagonal) matrix. We can for example choose K as the identity matrix multiplied by a factor k1, which is tuned by the user. Then, we can simply compute a solution using the same pseudo-inverse technique as described above [see Equation (24) to this modified problem].

We described how we deal with the inverse kinematics of the end-effector position of a two-link arm but we can generalize this procedure to a n-link arm. The interested reader can refer to Appendix G.5.

3.2.3. Inverse kinematics: numerical simulation

Here we test the numerical method for inverse kinematics developed in section 3.2.2 on the four-dof two-link arm. Figure 4 shows one particular example of inverse kinematics for this two-link arm. First, we choose an end-effector motion with a bell-shaped velocity profile and a straight line path in the 3D workspace (see Figure 4B), to mimic hand velocity profiles described in the neuroscience literature (Abend et al., 1982; Atkeson and Hollerbach, 1985). From the subject perspective, the hand is moving from a central position to the right, up and away from the shoulder. The initial configuration of the arm is represented in black (see Figure 4A). From the end-effector motion, we apply the inverse kinematics technique described in the previous section to estimate the joint rotations over time for the movement (Figure 4C).


Figure 4. Numerical simulation example of inverse kinematics of the four-dof two-link arm. For this simulation, we used k0 = 1000 and k1 = 1 (see Appendix G.4) and a simulation step of 1 ms duration. (A) Top view (top), behind view (middle), and side view (bottom) of the computed inverse kinematics for the joints. The end-effector (point P) motion in 3D space is described by the blue line. The initial two-link arm configuration is represented in black. Upper arm (dark gray) and lower arm (light gray) are represented at different instants during the end-effector motion. (B) The end-effector speed (left) and displacement (right) are described as a function of time: the movement is a typical bell-shaped speed profile lasting 800 ms and with a peak velocity of 100 cm/s. The end-effector displacement is about 25 cm. (C) The estimated elbow (left) and shoulder (right) rotation angles are represented as a function of time. For the shoulder, there are three components of rotation: around the x-axis (red line), the y-axis (green line), and the z-axis (blue line). (D) Representation of the position error, the norm of the difference between the estimated and actual end-effector position, as a function of time. The maximal error is about 0.02 cm.

The estimation accuracy was assessed by computing the norm of the position error, i.e., the difference between the true end-effector position and the estimated end-effector position. The end-effector position was estimated by using the forward kinematic model (see Equation 19) and the estimated joint angles (and therefore rotation quaternions). We can observe (Figure 4D) that the position error was always smaller than 0.02 cm. This position error depends on the value of k1 (see Equation 26) that we use. For this simulation, we used k1 = 1000. The larger k1 the smaller is the position error, because we penalize the position error more with larger values of k1. But k1 can not be too large for discrete time systems. If k1 is too large, the system will be unstable and the error will grow. The threshold value for k1 depends on the simulation step Δt (see Appendix G.6 for details). For our simulation, we used Δt = 1 ms and therefore we could not choose a value for k1 larger than 2000.

We tested several directions in space for the end-effector motions and our inverse kinematics algorithm was successful in every case, inferring joint angles for the elbow and shoulder. Note that this method does not necessarily reproduce the joints angles observed in human subjects but it yields the solution with the smallest joint angle rotations.

3.3. Three-Link Arm Movements: Forward and Inverse Kinematics for End-Effector Position and Orientation

In this section, we describe the forward and inverse kinematics for the end-effector position and orientation of a three-link arm. The main difference compared to section 3.2 is that we perform line transformations (using dual quaternions) to transform the orientation.

For this section, we decided to provide an example where we directly use screw motion to model the joints (shoulder, elbow, and wrist) movements, but we could have kept the same approach as in section 3.2 by alternating rotations and translation (offsets). Similarly, we could have used screw motion dual quaternions as well in section 3.2 for the two-link arm example.

3.3.1. Forward kinematics for end-effector position and orientation

Here, we describe a seven-dof arm with three links (see Figure 5): the upper and lower arms as well as the wrist. We consider a seven-dof joint model: three for the shoulder, one for the elbow, and three for the wrist. Up to here, we could just apply the methodology described above in section 3.2. However, now we consider an example where the end-effector orientation matters. For example, Figure 5 shows that the hand holds a screwdriver, and we are interested in the location and orientation of this end-effector.


Figure 5. seven-dof three-link arm top view. The shoulder insertion point, elbow and wrist are represented as well as the end-effector (a screwdriver) location and orientation. The different reference frames are also represented: body-fixed shoulder-centered (BS, in red), upper arm -fixed shoulder-centered (US, in blue), lower arm -fixed elbow centered (LE, in green), and hand-fixed wrist-centered (HW, in magenta). The rotation angles of the shoulder, elbow, and wrist are respectively denoted φ, θ, and α. The configuration in gray represents the reference configuration.

In order to describe the end-effector (the screwdriver) orientation (in body-fixed coordinates), we now use the formalism of line dual quaternions (see section 2.3.2). Let L0=n0+ϵ(r0n0) be the line dual quaternion describing the reference line passing through the screwdriver in its reference configuration [see the gray configuration in Figure 5: the arm is fully extended, φ = θ = α = 0° and n0 = (0 1 0)T and r0 = (a + b + c 0 0)T]. The forward kinematics describing the end-effector line position, L(t) = n(t) + ϵm(t), is described by line transformations, using the tools described in the previous sections (again, here we use screw motions but we could have used rotations and translations offsets):

L(t)=(SUB*SLU*SHL*)L0(t)(SHLSLUSUB)Stot      =Stot*L0Stot(27)


SUB = RUB = cos(φ/2) + nUB sin(φ/2) is a pure rotation quaternion describing the 3D rotation of the upper arm around the shoulder.

SLU = T*ESRLUTES is a screw motion dual quaternion describing the 1D rotation of the lower arm around the elbow, whose rotation axis is offset from the origin (the shoulder):

RLU = cos(θ/2) + nLU sin(θ/2) is a pure rotation quaternion describing the 3D rotation of the lower arm around the elbow.

TES = 1 + ϵtes/2 where tes = [a 0 0]T is the offset between the elbow and the shoulder in the reference configuration.

SHL = T*WSRHLTWS is a screw motion dual quaternion describing the 3D rotation of the hand around the wrist, whose rotation axis is offset from the origin (the shoulder):

RHL = cos(α/2) + nHL sin(α/2) is a pure rotation quaternion describing the 3D rotation of the hand around the wrist.

TWS = 1 + ϵtws/2 where tws = [a + b 0 0]T is the offset between the wrist and the shoulder in the reference configuration.

We can also derive the expression for the end-effector line velocity, L˙(t), deriving the above expressions:

L˙(t)=Stot*L˙0(t)Stot       +SUB*SLU*(S˙HL*L0SHL+SHL*L0S˙HL)SLUSUB       +SUB*(S˙LU*LwSLU+SLU*LwS˙LU)SUB       +S˙UB*LeSUB+SUB*LeS˙UB(28)


  • Lw = S*HLL0SHL is the end-effector line position after applying the wrist rotation to the reference line L0.
  • Le = S*LULwSLU is the end-effector line position after applying the elbow rotation to the line Lw.

Remembering the description of a screw motion velocity dual quaternion (see Equation 13) in terms of the pure rotational component, Ω, the offset velocity term L and the translational velocity term along the screw axis , we can express these screw motion velocities for the seven-dof three-link arm. For each of the screw motion dual quaternions considered here, only the rotational part is moving, in which case simplifies to M˙=12TTL*RΩTL. Therefore, it is quite easy to compute L˙(t)=n˙(t)+ϵm˙(t).

From the line forward kinematics for position and velocity, we can derive the orientation of the end-effector, n(t), as well as its rate of change, (t). The actual position and velocity of the end-effector position is more tricky to obtain. Indeed, we can retrieve the quantities m(t) and (t), but with these quantities, we can only infer the component of the end-effector position orthogonal to the line orientation. And, even if the end-effector position is known, we can not uniquely determine the end-effector velocity component along the line orientation (see also Appendix B). Therefore, in parallel to the line transformation applied to the end-effector line, a position transformation should be applied to the end-effector position. The transformation equations applied to the reference point dual quaternion P0 = 1 + ϵr0 are similar to Equations (27) and (28) except that we use the conjugate definition S˜* instead of S* for point transformation (as described in section 2.4).

3.3.2. Inverse kinematics from end-effector position and orientation

In the case of line transformations, we are also interested in estimating the joint kinematic parameters from the knowledge of the end-effector position and orientation as well as the velocity and orientation rate of change. The problems and the challenges are similar to what we saw before in section 3.2.2 except for the construction of the jacobian matrix J which links the joint-velocities to the end-effector velocity and orientation rate of change. For the interested reader, Appendix H describes in detail how to compute this jacobian matrix J for the three-link arm and also generalizes to the n-link arm. Then we can proceed similarly to what is shown in section 3.2.2 to solve the inverse kinematics problem.

4. Discussion

In many applications of behavioral neuroscience or vision, there is a need to represent the 3D position and/or orientation of objects as well as their velocity. These objects may be external objects/stimuli that a human subject has to deal with (e.g., the position and orientation of a target object). For example, in order to catch the ball, a rugby player needs to estimate the 3D motion of the ball but also the time course of its orientation, since the ball is non-spherical. This object can also be a body part (e.g., the eye, the head, and the hand), whose position and/or orientation is of interest for the brain to monitor body movements. Dual quaternions provide a convenient compact and geometrically meaningful way to describe either position (through point dual quaternions) or position and orientation (through line dual quaternions). Moreover, dual quaternions provide a way to describe natural geometrical transformations like rotations, translations, or screw motions, and these geometrical transformations can easily be combined together and applied to points or lines. In this paper, we described the useful concepts of the dual quaternion geometric algebra and how dual quaternions can be used to model these transformations. We also described the dual quaternion formalism to cope with velocity: rotational velocity, screw motion velocity, point velocity, and line velocity. Then we applied these concepts to a few examples in behavioral neuroscience: the 3D eye-head-shoulder system reference frame transformation needed for the accurate planning of manual tracking movements. Another example was provided with the forward kinematics for the multi-link arm, either considering the end-effector (the hand) position alone, or considering the end-effector position and/or orientation. Finally we also derived the inverse kinematics of the same multi-link arm from the dual quaternion formalism. In these applications, we do not claim that the brain actually uses dual quaternions to implement these transformations. However, these complex transformations are easily expressed mathematically using dual quaternions, which helps the neuroscience researcher to make predictions, for a theoretical goal or just in designing an experiment.

A main advantage of dual quaternions is that we can combine several rotations and/or translations. Therefore, it is quite easy to compute and write the expression for complex systems. We used that advantage throughout our applications. Another advantage of using quaternions (or dual quaternions) to represent rotations is that it is an efficient way to parameterize rotations, without any singularity (Chou, 1992; Grassia, 1998), which occur when using the classical ways to parameterize rotations (Euler angles for instance). An important advantage of dual quaternions is the compactness in terms of memory requirements: we only need 8 elements to represent them, compared to 12 for homogeneous matrices (Kim and Kumar, 1990; Funda and Paul, 1990a; Aspragathos and Dimitros, 1998; Daniilidis, 1999; Kavan et al., 2008).

Quaternions have been widely used to model 3D eye rotations (Tweed and Vilis, 1987; Hestenes, 1994a; Haslwanter, 1995; Crawford and Guitton, 1997; Bayro-Corrochano, 2003) while dual quaternions have been used less (see Bayro-Corrochano, 2003; Blohm and Crawford, 2007; Leclercq et al., 2012). Here, we described how the dual quaternion formalism can be used to model the multi-body kinematics, which can be useful for modeling purposes in applications like motor planning for eye and arm movements, 3D eye-head gaze shifts, 3D VOR, 3D updating, computing predictions of an inverse or forward internal model.

Dual quaternions have been commonly used in robotics and computer vision, for various purposes. Quaternion parametrization of rotations are used in graphics applications (Grassia, 1998; Tolani, 2000; Azariadis, 2001; Kavan et al., 2008; Ding et al., 2012). Tolani (2000) developed a real-time inverse kinematic technique for anthropomorphic limbs. In Ding et al. (2012), they develop an image processing technique which is built upon a so-called dual quaternion Fourier transform. According to Grassia (1998), quaternions are the best choice to interpolate 3D rotations. Indeed, when we interpolate element by element between two rotations quaternions, it interpolates on the geodesic (shortest) path onto the sphere (Shoemake, 1985; Dam et al., 1998). Dual quaternions are also used for the process of pose estimation, which consists in estimating the position and orientation of an object. In this context, dual quaternions allow to solve simultaneously for both the position and orientation components (Daniilidis, 1999; Bayro-Corrochano et al., 2000) in the context of robotics (hand-eye calibration). Pose estimation is also used with dual quaternions in computer vision (Walker et al., 1991; Phong et al., 1993; Torsello et al., 2011), but with a projective component in addition. In Goddard and Abidi (1998), they used the iterative extended Kalman filter for this purpose of pose estimation in order to deal with uncertainty. Forward kinematics equations for robotic manipulators have been derived with dual quaternions (Kim and Kumar, 1990) and compared with other methods (Aspragathos and Dimitros, 1998). Perez and McCarthy (2004) use dual quaternions for the synthesis of constrained robotic systems where several serial constrained (less than six-dof) manipulators are combined. Inverse kinematics techniques have been developed using dual quaternions for six-dof manipulators (Aydin and Kucuk, 2006; Sariyildiz and Temeltas, 2009) and compared to other methods (Sariyildiz et al., 2011). They are also used in the context of cooperative control of multiple manipulators: several robotic manipulators (Adorno et al., 2010) or a robotic manipulator interacting with a human arm (Adorno et al., 2011). In Pham et al. (2010), they develop a control law using dual quaternion to control simultaneously the position and orientation of a robotic manipulator.

Thus dual quaternions are powerful mathematical constructs that are widely used in robotics and computer vision, and could make important contributions to neuroscience research.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.


This paper presents research results of the Belgian Network DYSCO (Dynamical Systems, Control, and Optimization), funded by the Interuniversity Attraction Poles Programme, initiated by the Belgian State Science Policy Office. The scientific responsibility rests with its author(s). This work was also supported by the Belgian Fonds National de la Recherche Scientifique, the Fondation pour la Recherche Scientifique Médicale, and internal research grants from the Université catholique de Louvain (Fonds Spéciaux de Recherche, Action de Recherche Concertée), The Botterell Foundation (Queen's University, Canada), NSERC (Canada), and the Ontario Research Fund (Canada).


Abend, W., Bizzi, E., and Morasso, P. (1982). Human arm trajectory formation. Brain 105, 331–348.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Adorno, B., Fraisse, P., and Druon, S. (2010). “Dual position control strategies using the cooperative dual task-space framework,” in Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference, (Taipei), 3955–3960.

Adorno, B. V., Bo, A. P. L., and Fraisse, P. (2011). “Interactive manipulation between a human and a humanoid: when robots control human arm motion,” in 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, (San Fransisco, CA), 4658–4663

Aspragathos, N. A., and Dimitros, J. K. (1998). A comparative study of three methods for robot kinematics. IEEE Trans. Syst. Man Cyberne. B Cybern. 28, 135–145.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Atkeson, C. G., and Hollerbach, J. M. (1985). Kinematic features of unrestrained vertical arm movements. J. Neurosci. 5, 2318–2330.

Pubmed Abstract | Pubmed Full Text

Atkinson, K. (1989). An introduction to numerical analysis, 2nd Edn. New York, NY: John Wiley Edition.

Aydin, Y., and Kucuk, S. (2006). “Quaternion based inverse kinematics for industrial robot manipulators with Euler Wrist,” in 2006 IEEE International Conference on Mechatronics, (Budapest), 581–586.

Azariadis, P. (2001). Computer graphics representation and transformation of geometric entities using dual unit vectors and line transformations. Comput. Graph. 25, 195–209.

Bayro-Corrochano, E. (2003). Modeling the 3D kinematics of the eye in the geometric algebra framework. Pattern Recogn. 36, 2993–3012.

Bayro-Corrochano, E., Daniilidis, K., and Sommer, G. (2000). Motor algebra for 3D kinematics: the case of the hand-eye calibration. J. Math. Imaging Vis. 13, 79–100.

Blohm, G. (2012). Simulating the cortical 3D visuomotor transformation of reach depth. PLoS ONE 7:e41241. doi: 10.1371/journal.pone.0041241

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Blohm, G., and Crawford, J. D. (2007). Computations for geometrically accurate visually guided reaching in 3-D space. J. Vis. 7, 4.1–4.22.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Blohm, G., Keith, G. P., and Crawford, J. D. (2009). Decoding the cortical transformations for visually guided reaching in 3D space. Cereb. Cortex 19, 1372–1393.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Blohm, G., Khan, A. Z., Ren, L., Schreiber, K. M., and Crawford, J. D. (2008). Depth estimation from retinal disparity requires eye and head orientation signals. J. Vis. 8, 3.1–3.23.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Cheng, H., and Gupta, K. (1991). A study of robot inverse kinematics based upon the solution of differential equations. J. Robot. Syst. 8, 159–175.

Chou, J. (1992). Quaternion kinematic and dynamic differential equations. IEEE Trans. Robot. Automat. 8, 53–64.

Crawford, J. D., and Guitton, D. (1997). Visual-motor transformations required for accurate and kinematically correct saccades. J. Neurophysiol. 78, 1447–1467.

Pubmed Abstract | Pubmed Full Text

Dam, E., Koch, M., and Lillholm, M. (1998). Quaternions, interpolation and animation. Technical report, University of Copenhagen.

Daniilidis, K. (1999). Hand-eye calibration using dual quaternions. Int. J. Robot. Res. 18, 286–298.

Ding, Z., Yu, Y., Wang, B., and Zhang, L. (2012). An approach for visual attention based on biquaternion and its application for ship detection in multispectral imagery. Neurocomputing 76, 9–17.

Funda, J., and Paul, R. P. (1990a). A computational analysis of screw transformations for robotics. IEEE Trans. Robot. Automat. 6, 348–356.

Funda, J., and Paul, R. P. (1990b). A computational analysis of screw transformations in robotics. IEEE Trans. Robot. 6, 348–356.

Funda, J., Taylor, R. H., and Paul, R. P. (1990). On homogeneous transforms, quaternions, and computational efficiency. IEEE Trans. Robot. Automat. 6, 382–388.

Goddard, J. S., and Abidi, M. A. (1998). “Pose and motion estimation using dual quaternion-based extended Kalman filtering,” in Proceedings of SPIE Conference on Three-Dimensional Image Capture and Applications. Vol. 3313, (San Jose, CA), 189–200.

Grassia, F. (1998). Practical parameterization of rotations using the exponential map. J. Graph. Tools 3, 29–48.

Haslwanter, T. (1995). Mathematics of three-dimensional eye rotations. Vis. Res. 35, 1727–1739.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Hestenes, D. (1986). New Foundations for Classical Mechanics. Dordrecht: D. Reidel Publishing Company.

Hestenes, D. (1994a). Invariant body kinematics: I. Saccadic and compensatory eye movements. Neural Netw. 7, 65–77.

Hestenes, D. (1994b). Invariant body kinematics: II. Reaching and neurogeometry. Neural Netw. 7, 79–88.

Hore, J., Watts, S., and Vilis, T. (1992). Constraints on arm position when pointing in three dimensions: Donders' law and the Fick gimbal strategy. J. Neurophysiol. 68, 374–383.

Pubmed Abstract | Pubmed Full Text

Kavan, L., Collins, S., Žára, J., and O'Sullivan, C. (2008). Geometric skinning with approximate dual quaternion blending. ACM Trans. Graph. 27, 1–23.

Kim, J.-H., and Kumar, V. R. (1990). Kinematic of robot manipulators via line transformations. J. Robot. Syst. 7, 649–674.

Klein, C., and Huang, C. (1983). Review of pseudoinverse control for use with kinematically redundant manipulators. IEEE Trans. Syst. Man Cybern. SMC-13, 245–250.

Leclercq, G., Blohm, G., and Lefèvre, P. (2012). Accurate planning of manual tracking requires a 3D visuomotor transformation of velocity signals. J. Vis. 12, 1–21.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

McGuire, L. M. M., and Sabes, P. N. (2009). Sensory transformations and the use of multiple reference frames for reach planning. Nat. Neurosci. 12, 1056–1061.

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Perez, A., and McCarthy, J. M. (2004). Dual quaternion synthesis of constrained robotic systems. J. Mech. Des. 126, 425.

Pham, H., Perdereau, V., Adorno, B., and Fraisse, P. (2010). “Position and orientation control of robot manipulators using dual quaternion feedback,” in Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference, (Taipei), 658–663.

Phong, T., Horaud, R., Yassine, A., and Pham, D. (1993). “Optimal estimation of object pose from a single perspective view,” in 1993 (4th) International Conference on Computer Vision (Berlin: IEEE Computer Society Press), 534–539.

Sariyildiz, E., Cakiray, E., and Temeltas, H. (2011). A comparative study of three inverse kinematic methods of serial industrial robot manipulators in the screw theory framework. Int. J. Adv. Robot. Syst. 8, 9–24.

Sariyildiz, E., and Temeltas, H. (2009). “Solution of inverse kinematic problem for serial robot using dual quaterninons and plücker coordinates,” in 2009 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, (Singapore), 338–343.

Sciavicco, L., and Siciliano, B. (1996). Modeling and Control of Robot Manipulators. New York, NY: McGraw-Hill.

Shadmehr, R., and Wise, S. P. (2005). The Computational Neurobiology of Reaching and Pointing: A Foundation for Motor Learning. Computational Neuroscience. Cambridge, MA: MIT Press.

Shoemake, K. (1985). Animating rotation with quaternion curves. ACM SIGGRAPH Comput. Graph. 19, 245–254.

Straumann, D., Haslwanter, T., Hepp-Reymond, M.-C., and Hepp, K. (1991). Listing's law for eye, head and arm movements and their synergistic control. Exp. Brain Res. 86, 209–215.

Pubmed Abstract | Pubmed Full Text

Tolani, D. (2000). Real-time inverse kinematics techniques for anthropomorphic limbs. Graph. Models 62, 353–388.

Pubmed Abstract | Pubmed Full Text

Torsello, A., Rodola, E., and Albarelli, A. (2011). “Multiview registration via graph diffusion of dual quaternions,” in CVPR 2011 (IEEE), (Providence, RI), 2441–2448.

Tweed, D., and Vilis, T. (1987). Implications of rotational kinematics for the oculomotor system in three dimensions. J. Neurophysiol. 58, 832–849.

Pubmed Abstract | Pubmed Full Text

Walker, M. W., Shao, L., and Volz, R. A. (1991). Estimating 3-D location parameters using dual number quaternions. CVGIP Image Underst. 54, 358–367.

Wang, L., and Chen, C. (1991). A combined optimization method for solving the inverse kinematics problems of mechanical manipulators. Robot. Automat. IEEE Trans. 7, 489–499.


A. Multiplication of Two Quaternions: Proof

The multiplication of two quaternions A=a0+ia and B=b0+ib is a quaternion C=c0+ic. It is computed as follows:

C=AB  =(a0+ia)(b0+ib)  =a0b0+i(a0b+b0a)ab  =a0b0+i(a0b+b0a)(a·b+ab)  =(a0b0a·b)c0+i(a0b+b0aa×b)c

using the property i2 = −1 and the expression of the geometric product of two vectors (see Equation 1).

B. Proof that the Line Derivative L˙ does not Depend on the Choice of p

In section 2.3.2, we saw that:

L˙=n.+ϵm.  =n.+ϵ(p˙n+pn˙)(B.1)(B.2)

One could argue that this line derivative expression depends on the choice of p to describe a point on line L, and how it changes on this line. We prove in the following that it is not the case. Assume we choose a particular representation: p1 and p1.. Then the dual component of expression (B.2) is:


Assume now that we choose another representation, p2 and p2. and that they are linked to the first one in the following way:


Then, the dual component of expression (B.2) is:

c2=p2.n+p2n˙   =(p1.+x˙n+xn˙)n+(p1+xn)n˙   =p1.n+x˙nn=0+xn˙n+p1n˙+xnn˙   =p1.n+p1n˙

which proves that c2 = c1 and that the representations are indeed independent of the choice of p.

C. Diverse Tools and Results

C.1. Shortest rotation between two unitary vectors

Let us consider two unitary vectors a1 and b1. Their quaternion representation (or dual part of their dual quaternion representation) are A1 = 0 + a1 and B1 = 0 + b1. We can write the following:

A1=B1(B11A1)    =B1(B1*A1)(C.1)

The expression B*1A1 can be developed:

B1*A1=b1a1       =b1a1       =b1·a1+b1a1       =cosθ+n sinθ(C.2)

where θ is the rotation angle from vector b1 toward vector b1, and n=b1a1sinθ is the unitary rotation axis. Together, these two parameters describe the shortest rotation from vector b1 to vector a1.

This way of finding the shortest rotation between two unitary vectors can for instance be used to describe the 3D eye rotation obeying Listing's law. As all 3D eye orientations can be described as a rotation from a reference eye position, the Listing's law states that all the possible rotation axes lie in a plane, which is orthogonal to a specific reference eye position, called the primary position. Knowing the primary position and the current gaze position, it is easy to compute the rotation quaternion: Rrot = cos θ + n sin θ describing the rotation angle and rotation axis of the eye, using Equation (C.2).

C.2. Shortest screw motion between two lines

Similarly to what is shown in Appendix C.1, we can find the shortest screw motion between two lines L0 = n0 + ϵm0 and L1 = n1 + ϵm1, where m0=r0n0 and m1=r1n1, r0 and r1 being arbitrary points belonging to lines L0 and L1. “Shortest” refers to the smallest rotation angle θ and translation distance d along the screw axis. Indeed, in 3D space, two lines do not usually intersect. d represents the shortest distance between the two lines, and is therefore non-zero when the lines do not intersect. We can write:

L1=L0(L01L1)   =L0(L0*L1S)

since L0 and L1 are unitary dual quaternions. It can be shown that S indeed represents the shortest screw motion transformation between L0 and L1 and has the following form:

S=L0*L1=cosθ+sinθnS+ϵ( dsinθ+dnScosθ        +(aSnS)sinθ )

Therefore, we can compute the angle θ and distance d between any two lines in space using this expression.

C.3. Screw motion velocity between two moving lines

Following what we did in the Appendix C.2, we can develop the derivatives of the screw motion dual quaternion S, to show how to extract quantities like the rotation angle derivative θ˙ or the screw motion translation derivative, d˙. Indeed, if we derive S, we directly obtain:

S˙=θ˙sinθ+(θ˙nScosθ+n.Ssinθ)+ϵ [ (d˙sinθ     +dθ˙cosθ)+( d˙nScosθ+dn.Scosθdθ˙nSsinθ     + (aSnS)θ˙cosθ+(aS.nS)sinθ+(aSnS.)sinθ ) ]

Although this expression is quite long and difficult to interpret, it is quite easy to retrieve θ˙ and d˙ from this dual quaternion, provided we retrieve the parameters of the screw motion S first.

D. Conversion from and to Fick Coordinates

In diverse applications, we might want to go from a rotation quaternion expression to Fick coordinates, or vice-versa. Here we briefly describe these mathematical transformations.

Fick coordinates are commonly used in the 3D eye literature (see Haslwanter, 1995), for more comprehensive explanations): they describe a 3D rotation by three successive rotations in a well-defined order. Illustrating the context of a 3D eye position, the first Fick coordinate is a horizontal rotation of θF around a head-fixed axis (θF is positive when the rotation is to the left). Then the second Fick coordinate is a vertical rotation of φF around the inter-aural axis (i.e., the spatial horizontal axis rotated by the first Fick rotation). φF is positive when the rotation is down. Finally, the third Fick coordinate is a torsional rotation of ψF around the line of sight (i.e., the spatial backward–forward axis rotated by the two first Fick rotations). ψF is positive when the rotation is clockwise. All the definitions given above are defined from the point of view of the object being rotated.

A rotation quaternion is described by a rotation angle θ and a rotation axis n and is written (see Equation 3):

Rrot=cos(θ2)+n sin(θ2)

D.1. Fick coordinates to rotation rotor

Here, we want to compute θ and n from Fick coordinates (θF, φF, ψF). First, we compute the 3 × 3 rotation matrix describing the rotation (see Haslwanter, 1995), using c and s as shortcuts for cos and sin:


Then, we need to compute the rotor Rrot = q0 + q from the rotation matrix. Funda et al. (1990) showed that if the rotation matrix R is written:


then, q0 and q = (q1, q2, q3) are given by:


This transformation is for instance needed to reconstruct the 3D eye rotation quaternion, from Fick coordinates provided by a video-based eye tracking system.

D.2. Rotation rotor to fick coordinates

Here we compute the Fick coordinates (θF, φF, ψF) from a rotation quaternion representation Rrot=q0+q=cosθ2+nsinθ2. From q0 and q = (q1, q2, q3), we can retrieve the rotation matrix R (see Funda et al., 1990):


Then, knowing the structure of R in Fick coordinates (Equation D.1), we can identify term by term and extract the Fick parameters (except when there are singularities, in which case one angle can have multiple values).

E. Derivation of PEE and PĖE

First we derive step by step the expressions for the pointing target P in eye coordinates. First we show that:


Indeed, we obtain this result by applying one operation (rotation or translation) at a time:


where PEH, PHH, and PHS are the coordinates of the pointing target in eye-centered head-fixed coordinates, head-centered head-fixed coordinates, and head-centered shoulder-fixed coordinates.

For the P velocity, we differentiate Equation (E.1). Differentiating the first line, we write:


Using the rotational velocity operator (Equation 4), Equation (E.2) may be written:


Because THE is time-invariant (the offset between eye and head rotation centers is obviously fixed in a head-fixed reference frame), we have then


Similarly to Equation (E.2), we have:


Finally, we assume THS is time-invariant. Thus,


The final expression for P velocity in an eye-centered eye-fixed reference frame is therefore:


F. Proof of ḊP Expression

Here we compute the expression of d˙P, which is the rate of change of the distance between the eye and the target P. Without loss of generality, we can write:




The distance between the eye and the target P, denoted dP, is computed from the expression of PEE:


Then, d˙P can be developed:

dP.=xx˙+yy˙+zz˙x2+y2+z2   =PEE·P.EEdP

G. Forward and Inverse Kinematics: Derivation of Results

G.1. Computation of the 2D velocity of the end-effector position

We show that:



PBS.=12(PBSnφ˙φ˙nPBS)+RUB*PUEnθ˙θ˙nPUE2RUB     =[ 12(PBSnnPBS) ]φ˙+[ RUB*PUEnnPUE2RUB ]θ˙     =(12(PBSnnPBS)RUB*PUEnnPUE2RUB)JDQ(φ˙θ˙)T

where n is the rotation axis of the shoulder and the elbow (since we only consider planar motions)

G.2. Proof that 12(PBSΩUBΩUBPBS)=ϵ(PBSΩUB)

First, using the definition of a point position dual quaternion PBS = 1 + ϵPBS, we develop the left side of this equation:

12(PBSΩUBΩUBPBS)=12[ (1+ϵPBS)ΩUBΩUB(1+ϵPBS) ]                                  =ϵ12(PBSΩUBΩUBPBS)

Then, let us compute the quaternion multiplications of the two bivectors PBS and ΩUB, using the quaternion properties developed in section 2.1:

PBSΩUB=(iPBS)(iΩUB)           =PBSΩUB           =PBS·ΩUBPBSΩUB

Similarly, we have:



12(PBSΩUBΩUBPBS)=ϵ12(PBSΩUBΩUBPBS)                                 =ϵ12(PBSΩUB(ΩUBPBS))                                 =ϵ12(PBSΩUB(PBSΩUB))                                 =ϵ(PBSΩUB)

G.3. Derivation of the matrix expression for ṖBS

Here, we start from Equation (19) which expresses the velocity of the point end-effector P as a function of the rotational velocities of the shoulder and elbow, using the dual quaternion formalism. As a reminder, this equation is:


where PUE is the position dual quaternion of point P in an upper-arm fixed elbow-centered reference frame:


Using the relationship derived in Appendix G.2,


we can develop Equation (G.1) as:


Using the rotation matrix RM corresponding to the rotation dual quaternion R [see Equation (D.2) in Appendix D for how to compute this rotation matrix from the rotation dual quaternion], the rotation of a point P, written as P′ = R*PR in dual quaternion formalism, is written as P=RMP in matrix notation. Furthermore, a cross product of two vectors v=a×b can also be written as v=A˜ab where


where a = (a1 a2 a3). Note that this matrix has always rank 2 since for an equation in x: a×x=v, the component of x parallel to a can be arbitrary. All the solutions x lie on a line.

Using these two properties, BS can be developed further and its vector velocity component, BS, can be written in matrix form as:


G.4. Computation of the pseudo-inverse J+

In this appendix, we provide a brief tutorial on the computation of the pseudo-inverse J + to the interested reader. We also explain how to deal with solutions which are close to singularities.

In practice, Equation (24) requires that we compute explicitly the pseudo-inverse J +, by applying for example an algorithm based on the svd (singular value decomposition) of J [see Klein and Huang (1983), it is also what Matlab does to compute it]. The advantage is that the pseudo-inverse exists whatever the rank k of the matrix J which is of size n × m, even if k < min(n, m). However, the computation time is quite large. If the matrix J has rank k = min(n, m), we can solve Jx=v˙ (where v˙ is the velocity of the end-effector) without explicitly computing the pseudo-inverse (Klein and Huang, 1983). Indeed, in the case where the system is underdetermined, k = m, m < n, the pseudo-inverse is:


while if the system is overdetermined, k = n, n < m, then:


Taking advantage of this formulation, if the system is underdetermined, we solve classically for:


where w is an intermediate variable from which x is computed by takingx = JT w. If the system is overdetermined, then the solution x is obtained by solving:


Using these methods, a solution for x is obtained much faster than by explicitly computing the pseudo-inverse J+. However, these methods do not work properly if there are singularities in the kinematic system. Singularities happen for example when the two links of the arm are aligned. In this case, the rank of the jacobian J is diminished by at least 1, and we can no longer apply this technique. Either, we manually remove one linearly dependent row of J, or we compute the explicit pseudo-inverse. Moreover, it can be that there are no solutions when a singularity is met. If the two links are aligned and if the velocity of the end-effector has a component aligned with the two links, there are no solutions (see Sciavicco and Siciliano, 1996).

Finally, in the vicinity of a singularity, the joint velocities may become very large. A solution to this last problem can be found by applying the damped least-squares (DLS) inverse instead of the pseudo-inverse. It consists in using the term JJT + k0I (where I is the corresponding identity matrix, and k0 is a scalar to be chosen) instead of JJT in Equation (G.2) or JTJ + k0I in Equation (G.3). The equations are therefore better conditioned. As k0 increases, the joint velocities decrease but at the price that they do not obey perfectly the equations Jx=v˙.

G.5. n-link arm: forward and inverse kinematics of the end-effector position

In the main text, we have introduced the problem of inverse kinematics and its various difficulties through the example of the four-dof two-link arm: how to deal with redundancy, how to deal with singularities, how to deal with imperfect numerical integration. In the following, we describe how to build the jacobian matrix for a general serial manipulator (something like an arm with n links). The methodology and the problems are the same but the size of the system is bigger.

The forward kinematic of a n-link arm end-effector position is:

PN=(RN*TN1RN1*T1R1*)P0(R1T1RN1TN1RN)     =RN*(i=1N1TiRi*)S˜tot*P0(i=1N1RiTi)RNStot

where P0 is the end-effector position dual quaternion in the end-effector fixed reference frame, Ri=cosθi2+nisinθi2 is the rotation quaternion representing the rotation from link Ni + 1 to link Ni (link 0 is the inertial base reference frame) in Ni link-fixed coordinates, and Ti=1+ϵdi2ti is the translation from link Ni + 1 to link Ni (link 0 is the inertial base reference frame) in Ni link-fixed coordinates.

Its velocity can be derived:

PN.=(j=1NRj)*P0.(j=1NRj)        +i=1N1[ (j=i+1NRj)*2Ti.(j=i+1NRj) ]        +i=1N[ (j=i+1N+1Rj)*PiΩiΩiPi2(j=i+1N+1Rj) ]

where RN + 1 = 1, which is set in order to avoid writing one additional line for the term of the sum where i = N, and where

Pi=[ (j=1i1RjTj)Ri ]*P0[ (j=1i1RjTj)Ri ]

where i is the translation velocity dual quaternion from link Ni + 1 to link Ni, Ωi is the rotational velocity between link Ni + 1 to link Ni, and 0 is the end-effector velocity dual quaternion in the end-effector fixed reference frame (most of the time, 0 = 0). From this expression, we can build the jacobian J and then write the inverse kinematic equations Jx=v˙:


Then, we predict the rotation and translation dual quaternions at time t + Δt, Rj(t + Δt), and Tj(t + Δt), similarly to Equation (25).

G.6. Threshold value for the feedback term in the numerical procedure

As explained in section 3.2.3 where we developed the numerical simulation example, we can take into account the difference between the true end-effector position and the end-effector position estimated through the reconstructed joint angles. To do so, we tune a parameter, k1. The larger k1 the smaller is the position error, because we penalize the position error more with larger values of k1. But k1 can not be too large for discrete time systems. If k1 is too large, the system will be unstable and the error will grow. The threshold value for k1 depends on the simulation step Δt. Indeed, the position error is described by a linear differential equation (see Sciavicco and Siciliano, 1996):


where e is the position error and K is the feedback matrix, that we fix for the following to be equal to k1 multiplied by the identity matrix I (we penalize the errors in all the dimensions similarly). In discrete time, we have:


which is reformulated as:

e[k+Δt]=(IKΔt)e[k]             =(1k1Δt)Ie[k]

From this expression, we see that 0<k1<2Δt for the system to be stable (in discrete time, the eigenvalues must lie inside the unit imaginary disk in order to be stable).

H. Inverse Kinematics from the End-Effector Position and Orientation

H.1. Three-link arm

In order to build the jacobian matrix, we need to transform Equation (28) [where x is the unknown and L˙(t) is known for our inverse kinematics application] into a matrix equation.

First, we need to use the matrix expression for a screw motion transformation applied to a line or a line velocity (which have the same structure). Let us consider a line L0 = n0 + ϵm0 to which we apply a general screw motion S: a rotation of angle θ around a line whose orientation is s and that is offset by a from the origin (it can be the position of any point on the line), followed by a translation d along the line axis s. Then, the expression:

L(t)=S*L0S=n(t)+ϵm(t) (H.1)

can be described by the following matrix expression (see proof in Appendix I.1):

(RM0              3×3A˜aRMRMA˜a+RMdA˜sRM)J(n0m0)=(nm)         (H.2)

where RM is the rotation matrix associated with the rotation quaternion cos(θ/2) + s sin(θ/2), 0 is the 3 × 3 matrix whose entries are all 0 and A˜a (resp. A˜s) is the anti-symmetric matrix associated with the cross product, a × · (resp. s × ·).

If we transform L0 through N screw motions S1, …, SN:


then the corresponding matrix expression can be computed by combining N matrices of the type of J in Equation (H.2):


Let us consider the expression of L(t) given by Equation (H.1). The line velocity, L˙(t) is given by:

L˙(t)=S˙*L0S+S*L0S˙+S*L0.S       =n.(t)+ϵm.(t)(H.5)

where L˙0=n˙0+ϵm˙0, S is the screw motion operator and its velocity. Let us write the matrix expression of this equation. The third term of Equation (H.5) consists simply of a screw motion transformation of the same form as shown in Equation (H.1), applied to the line velocity. The sum of the two first terms gathers the velocity component due to the screw motion velocity applied to the line L0. In Appendix I.2, we show that if we parameterize S as S = TT*LRTL as described in sections 2.5.1 and 2.5.2, then we have:

S˙*L0S+S*L0S˙=R*(ϵd˙(n0s))R                          +R*(ϵ(n0a˙))Rϵ((R*n0R)a˙)                          +TL*(Ωn1+ϵ(Ωm1))TL(H.6)

where n1 + ϵm1 = (TT*LR)*L0 (TT*LR). The first term of Equation (H.6) represents the influence of the translation velocity d˙ along the screw axis s onto the component of L˙. The second term represents the influence of the offset velocity ȧ between the screw motion line and the origin onto the component again. Finally, the last term represents the influence of the rotational velocity Ω around the screw motion line onto both the and components of the line velocity. In matrix terms, Equation (H.6) may be written:

(03×1   03×3   A˜n1RM(n0s)RMA˜n0A˜RMn0A˜aA˜n1A˜m1)K(d˙a˙Ω)=(n.m.)                (H.7)

From all these expressions, we can now build the matrix expression for Equation (28), knowing that for our application, L˙0=0 (the screwdriver end-effector has no intrinsic velocity in the reference configuration). Therefore we have:

(JUBJLUKHLJUBKLUKUB03×3   A˜nLU   03×3)A(ΩHLΩLUΩUB)=(n.m.03×1)      (H.8)

where A is the jacobian matrix of size 9 × 9 of rank 8, which is consistent with the fact that there are seven-dof and six parameters coding for the line L. In Equation (H.8), we have:

  • JUB, JLU, and JHL are matrices of the form J described in Equation (H.2). JUB has only a rotational component, while JLU and JHL also have an offset component for their rotation axis, aLU and aHL.
  • KUB, KLU, and KHL are matrices of the form of the last column of matrix K in Equation (H.7), since in the seven-dof three-link arm, we only consider rotational velocities. They are therefore of the type: Ki=(A˜niA˜aiA˜niA˜mi) for i = 1, 2, 3 where ni and mi are obtained as follows:

Then we can use the techniques described in the previous sections, using the jacobian matrix to solve the inverse kinematics problem. In the following, we generalize this method to a n−link arm whose end-effector orientation matters.

H.2. n-link arm

Again the end-effector reference orientation and position in the base frame is modeled by a line L0 = n0 + ϵm0. The N-link motions can each be described by a screw motion Si, i = 1, …, N. Therefore we have:


The line velocity can be expressed as:

L˙(t)=(i=1NSi)*L˙0(i=1NSi)         +i=1N[ (j=i+1NSj)*(S˙i*Li1Si+Si*Li1L˙i)(j=i+1NSj) ]

where Li=(j=1i1Sj)*L0(j=1i1Sj). Now, in matrix terms, using the properties that we derived above, we obtain:


where Ji is of the form described by Equation (H.2), using the rotational and translational parameters of link i:


and where Ki has the following structure:



  • (nimi)=(j=1iJj)(n0m0)
  • (nimi)=(RiM03×3RiMA˜aiRiM)(j=1i1Jj)(n0m0)

We can of course add constraints by including them into the jacobian matrix A of Equation (H.9), for example if we want to impose a rotation axis for Ωi. Then, using this jacobian matrix A, we can proceed like described before to find a solution for the inverse kinematics problem.

I. Matrix Expressions for Screw Motion Position and Velocity Transformations

I.1. Screw motion transformation

Here we want to express the following line transformation under a matrix form


where L0 = n0 + ϵm0. A screw motion S can be parameterized as S = TT*LRTL (see section 2.5.1), where R is a pure rotation dual quaternion, TL=1+ϵ12a represents the offset of the rotation axis from the origin and T = 1 + ϵds is the translation along the screw axis. Therefore, we write:


We can show that T*(n0 + ϵm0)T = n0 + ϵ(m0 + ds × n0). Performing one transformation, rotation or translation, at a time, we can write in matrix form:

(n(t)m(t))=(I3×3  03×3A˜aI  3×3)(RM  03×303×3  RM)(I3×3  03×3A˜aI  3×3)                  (I3×3  03×3dA˜sI  3×3)(n0m0)           =(RM    03×3A˜aRMRMA˜a+RMdA˜sRM)J(n0m0)

I.2. Screw motion velocity

Here we demonstrate that if we parameterize S as S = TT*LRTL as described in sections 2.5.1 and 2.5.2, then we have:

S˙*L0S+S*L0S˙=R*(ϵd˙(n0s))R                          +R*(ϵ(n0a˙))Rϵ((R*n0R)a˙)                          +TL*(Ωn1+ϵ(Ωm1))TL

where n1 + ϵm1 = (TT*LR)*L0 (TT*LR). Indeed, using Equation (13), we have:

S˙*L0S+S*L0S˙=(R*T˙*)L0(TTL*RTL)+(TL*R*TLT*)L0(T˙R)translationalvelocitycomponent  +(R*T˙L+T˙L*R*)L0(TTL*RTL)+(TL*R*TLT*)L0(T˙L*R+RT˙L)offsetvelocitycomponent  +(12TL*ΩR*TLT*)L0(TTL*RTL)+(TL*R*TLT*)L0(12TTL*RΩTL)rotationalvelocitycomponent

Using the following properties:

  • velocity dual quaternion due to translational velocity: (0+ϵv)T˙(n+ϵm)=ϵvnA
  • invariance of this velocity dual quaternion due to translations: (1+ϵa)TϵA=ϵA
  • rotation of this velocity dual quaternion due to translations: R*(0 + ϵA)R = ϵ(R*AR)

The velocity component due to the translational velocity along the screw axis can be developed as follows:

translationalvelocitycomponent          =(R*T˙*)L0(TTL*RTL)+(TL*R*TLT*)L0(T˙R)          =R*T˙*L0R+(TL*R*TLT*)L0(T˙R)          =R*[ ϵd˙2s(n0+ϵm0)+(n0+ϵm0)ϵd˙2s ]R          =R*[ ϵ(d˙2sn0+d˙2n0s ]R          =R*(ϵd˙(n0s))R

The velocity component due to the screw axis offset velocity can be expressed as:


Finally, the velocity component due to the rotational velocity is:

rotationalvelocitycomponent                   =(12TL*ΩR*TLT*)L0(TTL*RTL)                      +(TL*R*TLT*)L0(12TTL*RΩTL)                   =12TL*ΩL1TL+12TL*L1ΩTL                   =12TL*(L1ΩΩL1)TL                   =TL*(Ωn1+ϵ(Ωm1))TL

where L1 = n1 + ϵm1 = (R*TLT*)L0(TT*LR).

Keywords: geometric algebra, forward kinematics, inverse kinematics, reference frame transformation

Citation: Leclercq G, Lefèvre P and Blohm G (2013) 3D kinematics using dual quaternions: theory and applications in neuroscience. Front. Behav. Neurosci. 7:7. doi: 10.3389/fnbeh.2013.00007

Received: 31 July 2012; Accepted: 28 January 2013;
Published online: 20 February 2013.

Edited by:

Markus Lappe, Universität Münster, Germany

Reviewed by:

W. Pieter Medendorp, Radboud University Nijmegen, Netherlands
Thomas Haslwanter, Upper Austria University of Applied Sciences, Austria

Copyright © 2013 Leclercq, Lefèvre and Blohm. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in other forums, provided the original authors and source are credited and subject to any copyright notices concerning any third-party graphics etc.

*Correspondence: Guillaume Leclercq, Institute of Information and Communication Technologies, Electronics and Applied Mathematics, Université Catholique de Louvain, Bâtiment Euler, Avenue Georges Lemaitre, 4 1348 Louvain-la-Neuve, Belgium. e-mail:
Gunnar Blohm, Centre for Neuroscience Studies, Queen's University, Botterell Hall, Room 229, 18 Stuart Street, Kingston, ON, K7L 3N6, Canada. e-mail: