Sie sind auf Seite 1von 12

This article has been accepted for inclusion in a future issue of this journal.

Content is final as presented, with the exception of pagination.

IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS 1

Development of Sensory-Motor Fusion-Based


Manipulation and Grasping Control for a
Robotic Hand-Eye System
Yingbai Hu, Zhijun Li, Senior Member, IEEE, Guanglin Li, Senior Member, IEEE, Peijiang Yuan,
Chenguang Yang, Member, IEEE, and Rong Song

Abstract—In this paper, a sensory-motor fusion-based manip- I. I NTRODUCTION


ulation and grasping control strategy has been developed for
OBOTIC bionic arms and prosthesis hands are of great
a robotic hand-eye system. The proposed hierarchical control
architecture has three modules: 1) vision servoing; 2) sur-
face electromyography (sEMG)-based movement recognition; and
R use to help people with severe physical disabilities, which
could enable disable people to achieve greater independence
3) hybrid force and motion optimization for manipulation and and consequently increase quality of life [1]–[3]. In general,
grasping. A stereo camera is used to obtain the 3-D point cloud bionic arms and prosthesis hands can be controlled by a joy-
of a target object and provides the desired operational position.
stick, which enables an operator to perform manipulation,
The AdaBoost-based motion recognition is employed to discrim-
inate different movements based on sEMG of human upper open/close of fingers, and thus offers an intuitive interface.
limbs. The operational space motion planning for bionic arm However, the bionic arm control loop short of biological
and force planning for multifingered robotic hand can be both signals feedback makes it very tedious to perform simple
transformed as a convex optimization problem with various con- manipulations, e.g., safely picking up, holding, or releasing
straints. A neural dynamics optimization solution is proposed an object. Therefore, it is necessary to investigate the replace-
and implemented online. The proposed formulation can achieve
a substantial reduction of computational load. The actual imple- ment of the natural hand with a bionic arm and prosthesis hand
mentation includes a bionic arm with dextrous hand, high-speed using a close-to-natural human-machine interface. Humans are
active vision, and an EMG sensors. A series of manipulation extremely adept at controlling the high degree of freedom
tasks consisting of tracking/recogniting/grasping of an object hand-wrist-arm musculo-skeletal system and are able to grasp
are implemented, and experiment results exhibit the respon- and manipulate objects according to task requirements with
siveness and flexibility of the proposed sensory motion fusion
approach.
minimum effort. It is of great interest to study human manip-
ulation and grasping behavior in order to extract underlying
Index Terms—Bionic arm, multifingers hand, quadratic principles which may be transferred to robotic or prosthetic
programming, surface electromyography (sEMG), vision
servoing. devices.
Myoelectric control allows a convenient human-machine
interface by transforming electromyography (EMG) signals
into control outputs. Two control approaches are currently
Manuscript received October 2, 2015; revised January 27, 2016; accepted employed in conventional myoelectric control: 1) direct con-
March 23, 2016. This work was supported in part by the National Natural trol and 2) pattern recognition (PR). Direct control links
Science Foundation of China under Grant 61573147 and Grant 91520201, antagonistic muscles or muscle groups directly to a single
in part by the Guangzhou Research Collaborative Innovation Projects under
Grant 2014Y2-00507, in part by the Guangdong Science and Technology degree-of-freedom (DoF) [4], [20]. Various switching meth-
Research Collaborative Innovation Projects under Grant 2014B090901056, in ods have been presented for sequentially transition between
part by the Guangdong Science and Technology Plan Project (Application different DoFs or functions in finite state machines [5].
Technology Research Foundation) under Grant 2015B020233006, and in
part by the National High-Tech Research and Development Program These simplistic controllers provide reliable performance,
of China (863 Program) under Grant 2015AA042303. This paper was but lack the functionality to smoothly operate multiple
recommended by Associate Editor S. Nahavandi. (Corresponding author: DoFs [6].
Zhijun Li.)
Y. Hu and Z. Li are with the College of Automation Science and PR methods utilize machine learning techniques, including
Engineering, South China University of Technology, Guangzhou 510061, both classification [7] and regression [8], to decode a mapping
China (e-mail: zjli@ieee.org). between myoelectric inputs and desired outputs [21], which
G. Li is with the Shenzhen Institutes of Advanced Technology, Chinese
Academy of Sciences, Shenzhen 518055, China (e-mail: gl.li@siat.ac.cn). enhances functionality compared to direct controls by enabling
P. Yuan is with the Robotics Institute, Beihang University, Beijing 100191, multiple DoFs without explicit switching methods. However,
China (e-mail: itr@buaa.edu.cn). increased functionality requires an updated training set which
C. Yang is with the Zienkiewicz Centre for Computational Engineering,
Swansea University, Swansea SA1 8EN, U.K. (e-mail: cyang@theiet.org). is highly dependent on the users motion repeatability [7] and
R. Song is with the Department of Medical Engineering, Sun Yat-sen may be influenced by many external factors [9]–[11]. Thus,
University, Guangzhou 510275, China (e-mail: songrong@mail.sysu.edu.cn). the decoding often overfits to a small set of the full input
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org.
space, and consequently performance tends to degrade over
Digital Object Identifier 10.1109/TSMC.2016.2560530 time [12].
2168-2216 c 2016 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

2 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

An improvement method to the PR is combining concepts


of visual servoing (VS). These VS-based methods extend PR
principles to multiple DoFs through linear transformations
between sensor inputs and control outputs, naturally provid-
ing both simultaneous and proportional control. The mapping
creates a redundancy in the control scheme which reduces the
precision needed in muscle activations to control the entire
task-space.
VS is defined as the use of visual feedback mechanisms for
the kinematic control of a robot. Based on the positioning of
the camera on the link and control techniques, VS ramifies into
several types. Eye-in-hand and eye-to-hand VS are represented
by the position of the camera on the robotic manipulator.
Being attached on the robot arm, eye-in-hand VS provides Fig. 1. Mechanical design of the bionic arm and hand.
a narrower field of view as compared to eye-to-hand servo-
TABLE I
ing. Many control schemes use either a direct VS or a dual PARAMETERS OF B IONIC A RM AND D EXTEROUS H AND
loop system [19], [31]. A visual processing method for human
tracking using intelligent wheelchair was developed in [19],
and a VS for stabilization of an nonholonomic mobile robot
was proposed by incorporating a model predictive control [31].
An important issue in controlling a bionic arm combin-
ing a multifingered robotic hand for object manipulation and
grasping is the synthesis of the optimal motion planning and
contact force such that the planning trajectory and contact
forces are guaranteed under the stability of the manipula-
tion and grasping and its feasibility. This problem, known as
motion-force optimization (MFO) can be formulated as a con- transformed as a convex optimization problem with vari-
strained optimization problem. Due to computational time of ous constraints. A neural dynamics optimization solution
finding the solution, the MFO is usually performed offline. is proposed and suitable to be implemented online.
However, during the execution of a manipulation task, the The actual implementation utilizes a bionic arm with
planning trajectory of the arm and the position of the con- dextrous hand, high-speed active vision, and an EMG
tact points on the object, or the wrench (force and moment) sensors. A series of manipulation tasks consisting of track-
to be balanced by the contact forces, may change with time ing/recogniting/grasping of an object are implemented, and
and cannot be planned in advance. The planning trajectory and experiment results demonstrate the responsiveness and flex-
contact forces must be compatible with the various constraints ibility of the proposed sensory motion fusion approach.
depending on the joint velocity, the type of contact, as well
as on the joint torque limits of every joints. In these cases,
suitable algorithms for online computation of the solution of II. B IONIC A RM S YSTEM
the constrained MFO problem must be devised. The utilization The developed bionic arm is shown in Fig. 1, which consists
of recurrent neural networks is investigated in [22] and [23], of five revolute joints corresponding to shoulder abduction–
where the friction constraints and the joint torque limits were adduction, shoulder flexion–extension (flx-ext), elbow flx-ext,
considered. In [34], an approach was proposed for computing and wrist flx-ext, wrist pronation–supination (pron-sup). The
optimal grasping and manipulation forces, which utilizes neu- dc motor selected to activate robotic articulations was a Maxon
ral network to learn the nonlinear inverse kinematics functional dc flat brushless motor EC90 for the joint 1 and EC45f for
relating the hand joints positions and displacements to object other joints, respectively, and harmonic transmission drives are
displacement. used (model SHD-17-100-2SH for joints 1 and 3 and model
In this paper, a sensory-motor fusion-based manipulation SHD-14-100-2SH for joints 2, 4, and 5). The physical parame-
and grasping control for a robotic hand-eye system is devel- ters are listed in Table I. In the robotic device, high-resolution
oped. The proposed hierarchical control architecture has three encoders (1024 pulse/cycle) and Hall effect sensor are used to
modules: 1) vision servoing; 2) surface EMG (sEMG)-based measure the displacement angles between the joints.
movement recognition; and 3) hybrid force and motion opti- In Fig. 2, the robotic hand is approximately the same as
mization for manipulation and grasping. The contributions can human-hand in size and shape. It comprises of five fingers
be summarized as follows. (including one thumb) and a DOF in the thumb in the roll
1) The AdaBoost-based motion recognition is proposed to axis, which is suitable for stably grasping an object of arbitrary
discriminate different movements based on sEMG of the shape. Each finger except the thumb consists of two joints, the
multifingers. lower one is an active joint, the upper one is an passive joint.
2) The operational motion planning for bionic arm and The active joint is driven by a dc motor through worm gear
force planning for multifingered robotic hand is both through Maxon controller to control its motion (see Fig. 3).
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

HU et al.: DEVELOPMENT OF SENSORY-MOTOR FUSION-BASED MANIPULATION AND GRASPING CONTROL 3

Fig. 5. Front view of vision localization.

Fig. 2. Design of dexterous hand. (a) Whole view of the hand. (b) One
finger. (c) Inside view of hand chamber. (d) Thumb with 2 DOFs.

Fig. 6. Vertical view of vision localization.

III. V ISION F EEDBACK C ONTROL S YSTEM


A. Visual Sensor Preprocessing and Its Projection Model
The machine vision processing adopt a Point Grey
Bumblebee2 stereo camera with IEEE-1394 Firewire connec-
tion shown in Fig. 4. The Point Grey Bumblebee2 stereo
Fig. 3. Control architecture of the hand.
camera is a 2 sensor progressive scan CCD camera with fixed
alignment between the sensors. Video is captured at a rate of
20 frames/s with a resolution of 640 × 480 to produce dense
colored depth maps to assist in tracking and a viable pose
estimation of the object. The resolution-speed tradeoff has to
be managed concisely as an increased frame speed gives a
smooth robot trajectory whereas enhances the processing time.
And an increased resolution provides a denser, more accurate
point cloud for feature extraction but with increased latency.
In this paper, the adopted projection model is shown in Fig. 5.
Camera calibration is necessary as the use of lenses intro-
duces nonlinearities and deviates from the simple pin-hole
model such as lens distortion, namely radial and tangential dis-
tortion. The camera parameters, namely, the intrinsic, extrinsic
and distortion are evaluated by using a 2-D checker-board pat-
tern. 3-D reference models were avoided due to computational
Fig. 4. Structure of the vision sensor.
complexity and high cost of precise calibration objects. In this
paper, 20 checkerboard images were fed to the calibrator algo-
rithm encompassing differential angles in the projection space.
The passive joint is driven by a linkage, its joint rotation range
This provides enough values to estimate the camera geometry
is limited by the active joint. The motion of the passive joint
parameters [18].
derives from the natural motion of human hand. The palm is
constructed with five aluminum plates connected with screws.
The finger can be mounted on the palm directly. Through the B. Three-Dimensional Reconstruction
four holes located at the rear plate in the palm, the hand can The images captured by the Bumblebee2 stereo camera in
be attached to the bionic arm. active ambient lighting are shown in Fig. 7. Both images are
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

4 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

Fig. 7. Object landmark.

calibrated using the camera intrinsics and corrected for distor-


tion. Subsequently, the undistorted images are stereo rectified
in order to align the epipolar lines of both projection planes
and ensure the presence of similar pixels in a specified row
of the image. The images obtained are then frontal parallel Fig. 8. Schematic of the bionic hand grasping based on sEMG.
and ready for correspondence estimate. The essential and the
fundamental matrix are calculated by using epipolar geometry.
The essential matrix is a 3×3 matrix with five parameters; two choose the red color as a sign of object. The image is con-
for translation and three for the rotation values between the verted into L × a × b color space and the Euclidean distance
camera projection planes. On the other hand, the fundamental between red-green and yellow-blue opponent components of
matrix represents the pixel relations between the two images the object and a and b matrices calculated [27], [28]. Here,
and has seven parameters, two for each epipole and three for represents the L∗ luminance component, while a∗ and b∗ repre-
homography that relates the two image planes. Bouguets algo- sent color components. The formulas for converting an RGB
rithm is then implemented to align the epipolar lines and shift image into the coordinates can be found in previous works
the epipoles to infinity. (e.g., [18] and [19]). The minimum value gives the most accu-
Stereo correspondence is a method of matching pixels rate estimate of the object. In the color space, the Euclidean
with similar texture across two co-planar image planes. distance between (L1 ∗ , a1 ∗ , b1 ∗ ) and (L2 ∗ , a2 ∗ , b2 ∗ ) can be
The distance between the columns of these perfectly matched defined as

pixels is defined as d = xl − xr , where xl is the column value  2
Eab = (L2 ∗ − L1 ∗ )2 + (a2 ∗ − a1 ∗ )2 + b2 ∗ − b1 ∗ (3)
of left image pixel, and xr is the column value of right image
pixel. The vertical view is shown in Fig. 6. which is approximately equivalent to the perceptual difference
Block matching is implemented to evaluate the correspon- between these two colors. Furthermore, the corners of the red-
dence between the images. Block sizes of 15-pixel window object are calculated by Harris corner detector and the centroid
are used to find the matches by the use of sum of absolute calculated by intersection of the diagonals. The depth value of
differences. A semiglobal method is used to force the dispar- the centroid is then extracted from reconstructed point cloud
ity values to the neighboring pixels for a more comprehensive of the task space.
result. Disparity is inversely proportional to the depth of the
pixel and related by the triangulation equation IV. EMG B IO -F EEDBACK S ERVO S TRUCTURE
f In this section, in order to perform the grasping of dexter-
D=T (1) ous hand, we employ EMG signal as control feedback and
d
the real-time EMG motion recognition can be implemented
where f is the focal length, d is the disparity, and T is the
for controlling the dexterous hand [20], [26]. Fig. 8 shows the
baseline.
schematic diagram of grasping manipulation using EMG sig-
Triangulation refers to the estimation of depth of an object
nals. The fundamental components are the EMG acquisition,
by visualizing its location from two different known points.
feature extraction unit, and the classifier unit. The adopted
The reconstruction of the image in the Cartesian coordinates
classification is AdaBoostSVM.
is obtained by the use of projection matrix evaluated using
Bouguets algorithm. The 3-D reconstruction of the robots
workspace A. Data Acquisition
 T  T There are many types hand/wrist motions when studying the
Q x, y, d, 1 = X, Y, Z, W (2) dextrous hand. In daily life, all the hand motion can be sim-
plified as three types: 1) hand closing; 2) hand opening; and
where D is the depth, Q is the projection matrix, and X, Y, Z,
3) finger pinching, shown in Fig. 9. Four surface electrodes
and W is the coordinates.
were used to acquire the sEMG signals from the extensor dig-
itorum, flexor digitorum superficialus, extensor carpiradialis,
C. Object Detection and flexor carpiulnaris, respectively, which are related with the
In this paper, color-based segmentation is used in order three motions. In order to collected high quality sEMG signals,
to separate a single color object from the background, we the skin was scrubbed with alcohol and shaved if necessary,
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

HU et al.: DEVELOPMENT OF SENSORY-MOTOR FUSION-BASED MANIPULATION AND GRASPING CONTROL 5

Algorithm 1 Process of AdaBoostSVM


1.Input: The training samples are labeled as
(x1 , y1 ), ..., (xN , yN ); select the initial σ , σini , the minimal
σ , σmin , the step of σ , and σstep .
2.Initialize: the weights of training samples are chosen as
: w1i = 1/N, for all i = 1, . . . , N.
3.Do While(σ <σmin )
Fig. 9. Three postures of the grasping hand. (1) A RBFSVM component classifier is trained, ht , on the
weighted training set. 
TABLE II (2) The training error of ht is chosen : t = N i=1 wi , yi  =
t
D EFINITION OF F EATURES
ht (xi );
(3) If t > 0.5, choose σ value as σstep and goto (1).
(4) Choose weight for the component classifier ht : αt =
1−t
2 ln( t )
1

(5) Update the weights of training samples: wt+1 i =


wti exp(−αt yi ht (xi ))
, i = 1,...,N.
Ct 
4.Output: f(x) = sign( Tt=1 αt ht (x)).

and then the electrodes with conductive gel were attached to


the corresponding positions. better generalization using a fixed (optimal) σ value, whose
In this paper, four groups of sEMG signals were collected model parameters are adaptively different.
from one intact-limb subjects forearm. The four groups were However, AdaBoostSVM updates the weights of training
used for the training section. We collected sEMG signals of samples through the called reweighing technique. At initial-
each hand and each posture constitutes a subset. The subject ization, a large value is selected for σ , which is similar to
is asked to perform the three gestures. The subject holds his an RBFSVM classifier with very weak learning ability. Then,
posture 7 s and repeats seven times. RBFSVM with σ is trained as many steps until more than half
accuracy can be achieved. Otherwise, we can regulate the σ
value decreased slightly such that the learning capability of
B. Feature Extraction
RBFSVM is increased to enhance more than half accuracy.
In order to control a robotic hand in real time without per- The accuracy of new RBFSVM would not be too strong for
ceiving a time delay, the processing time of sEMG PR should the current weighted training samples through decreasing the
be less than 300 ms. Thus, the scheme of a sliding window σ value slightly, and thus we can obtain moderately accurate
with an incremental window is adopted for the steady-state RBFSVM component classifiers. These larger diversities may
motion recognition. For the real-time myoelectric hand con- lead to a better generalization performance of AdaBoost. This
trol, all the processes including transmitting control commands process continues until the σ is decreased to the given minimal
should be completed within an incremental window. In this value. The training process can be found in Algorithm 1.
paper, a 64 ms (64 samples) sliding window with a 16 ms
(16 samples) incremental window is selected. After data seg-
V. H YBRID M OTION AND F ORCE O PTIMIZATION
mentation, 14 features would be extracted in a sliding window.
In this paper, we choose the feature set which includes: mean A. Kinematics Redundant Resolution Formulation
absolute value (MAV), waveform length (WLEN), zero cross- Consider a bionic arm and denote the transformation of joint
ings (ZCs), and slope sign changes (SSCs). The definition of space vector q ∈ Rn to end-effector’s relative position vector
these features is listed in Table II. We can know that each vec- χ ∈ R with q̇ = [q̇T1 , . . . , q̇Tn ]T ∈ Rn , where q̇i is the joint
tor of features is 8-dimensional, which consists of four features velocity. The kinematic redundancy resolution is to find the
on each signal channel. set of joint velocity, which are feasible with respect to the
kinematic structure of the bionic arm, the corresponding joint
C. Feature Classification Based on AdaBoostSVM position and velocity limits, and minimize the joint velocity.
It is well known that the weak classifiers employed in The balance equation for the generalized forces applied to the
AdaBoost are support vector machines (SVMs) with the RBF object can be written in the form
kernel, and SVM algorithm is of a very high accuracy in PR. J(q)q̇ = χ̇ (4)
However, the proposed AdaBoostSVM has been proven with
better generalization performance than SVM on imbalanced where J(q) ∈ Rn×ρ is the Jacobian matrix.
classification problems [21]. In this paper, the constraints of position and velocity can be
Traditional RBFSVM component classifiers choose a large represented by
σ value (implying weak learning) on the initialization, the σ
value would converge as the boosting iteration proceeds. It qmin ≤ q(t) ≤ qmax (5)
would lead to a set of RBFSVM component classifiers with q̇min ≤ q̇(t) ≤ q̇max (6)
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

6 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

where qmin and qmax are the boundedness of the joint position, Therefore, the grasping force optimization problem can be
and q̇min and q̇max are the boundedness of the joint velocity. formulated as finding the optimal grasp wrench minimizing
The inequalities (5) and (6) can also be described by the the internal forces acting on the object, under the above con-
positive definite matrix straints. The internal forces are contact wrenches that satisfy
the friction cone constraints and belong to the null space of
K(q, q̇) = diag(P, V) > 0 (7) the grasp matrix G.
where The frictional inequalities (14) can be described by
 
PB,L q − qmin  
P= = (8) F(c) = diag F1 ( f1 ), . . . , Fn ( fn ) > 0 (18)
PB,H −q + qmax
 
VB,L q̇ − q̇min where Fi ( fi ) is the symmetric (2 × 2) matrix
V= = (9)
VB,H −q̇ + q̇max

fi,x fi,y

which contains the kinematic constraints from the lower limits fi,z + μi μi
Fi (ci ) = ci,y fi,x . (19)
PB,L , VB,L and upper limits PB,H , VB,H , respectively. fi,z −
μi μi
Hence, we can denote the simultaneous satisfaction of the
constraints of both position and joint velocity as Similarly, the torque limit constraint (15) can also be
K(q, q̇) = diag(P, V) > 0. (10) described by

The optimization problem is defined based on the minimiza- T( f , q, τe ) = diag(τB ) > 0 (20)
tion of the cost function (q, q̇) as
where
(q, q̇) = q̇T Qq̇/2 + bT (q(t) − q(0)) (11)
 
subject to J(q)x = χ̇ (12) τ J T (q)f − τL + τe
τB = B,L = (21)
ξmin ≤ K(q, q̇) ≤ ξmax (13) τB,H −J T (q)f + τh − τe

where Q is symmetric positive definite matrix, and ξmin and with the lower limit (τB,L ) and upper limit (τB,H ), respectively.
ξmax are the boundedness of the position and velocity. Hence, we can denote the simultaneous satisfaction of
both frictional and joint torque constraints as the posi-
B. Grasping-Force Optimization Formulation tive definiteness of the linearly constrained block-diagonal
Consider an object grasped by a multifingered robotic matrix
hand, there exist l contact points between the object and
the fingertips, the contact wrench of the grasp is denoted by T = diag(F, T) > 0. (22)
f = [ f1T , . . . , flT ]T ∈ Rlk with the ith contact vector fi ∈ Rk in
dimension k, and the generalized external force acting on the Let f (F) be the contact wrench vector extracted from the
object is denoted by fe ∈ R6 . frictional constraint matrix, and τB (T) as the vector composed
In this paper, we assume the point contact model with by the diagonal elements of T, and ζ (T ) = [f (F)T , τB (T)T ]T ,
friction is adopted and the joint actuators could provide the the linear constraints on matrix T can be described as
required torques. The contact wrench consists of three vec-
tors: the normal component fi,z to the object surface, and the Aζ (T ) = b (23)
two components fi,x , fi,y on the tangent plane. The friction and
joint torque constraint can be described by with
⎡ ⎤ ⎡ ⎤
1 2 G 0 0 fe
f + fi,y
2 ≤f (14)
A = ⎣ J T (q) 0⎦ b = ⎣ τL − τe ⎦
i,z
μi i,x −I
τL ≤ τ ≤ τU (15) J T (q) 0 I τH − τe
with the tangential friction coefficient μi at the ith con- the null matrix is 0 and the identity matrix of proper dimen-
tact point, the lower τL and upper joint torque bound τU , sion is I.
respectively. Through minimizing the cost function (T ), the optimiza-
Then, we can obtain the following relationship as: tion can be defined as
fe = Gf (16)  t
J T (q)f + τe = τ (17) (T ) = T QT + b
T T
T (t)dt (24)
0
where the grasping matrix G ∈ R6×lk is full rank for force- subject to Aζ (P) = b (25)
closure grasps, the external torque τe includes gravity, Coriolis, ξmin ≤ T ≤ ξmax (26)
centripetal and inertia effects at the fingers joints, τ is the
actuator torque, and J(q) ∈ Rlk×p is the hand Jacobian matrix where Q is symmetric positive definite matrix, and ξmin and
with the total number of the hand joints p. ξmax are the boundedness of T .
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

HU et al.: DEVELOPMENT OF SENSORY-MOTOR FUSION-BASED MANIPULATION AND GRASPING CONTROL 7

C. Neuro-Dynamics Optimization where γ > 0 is designed parameters, which is used to scale


Considering the motion optimization (11)–(13) and force the convergence rate, and E ∈ R(n+m)×(n+m) is an identity
optimization (24)–(26), the following constrained hybrid MFO matrix.
scheme for bionic arm can be summarized as Theorem 1: Consider any initial states, the state vector
 t [X ∗ , y∗ ]T of the recurrent neural network (33) converges to
minimize X QX + λ
T
X(t)dt (27) an equilibrium point [X ∗ , y∗ ]T , of which the first n elements
0 constitute the optimal solution X ∗ to the original QP prob-
subject to J X = ϒ (28)
lem (27)–(29). The exponential stability can be achieved,
− +
ξ ≤X≤ξ (29) provided there exists a constant σ > 0 such that
where X ∈ Rn , Q ∈ Rn×n is a symmetric positive definite    
matrix, λ > 0, ξ − ∈ Rn , and ξ + ∈ Rn are the lower and upper  X X I −J T X
 − P − ·
limits of the a joint variable vector for the motion optimization  y 
y J 0 y
 t 
or grasping for force optimization, respectively, and J is the λ 0 X(s)ds 2
+ 
corresponding Jacobian matrix. −ϒ 2
  ∗ 
The dual decision variables can help to solve quadratic pro-  X X 
≥ σ 2
gramming problem (27)–(29) effectively with duality theory.  y − y∗  2 . (34)
First, according to the Lagrange method [22], [23], defining ω
as the m-dimensional ∞ numerically and y ∈ Rm as the dual
Proof: A Lyapunov function candidate can be defined as,
decision vector defined for (28).
where the · 22 denotes the square of Euclidean two-norm
Quadratic program (27)–(29) merged with a bound con-
straint can be defined as the linear variational inequalities     ∗ 
X  X X 
= 2
 y − y∗  2 ≥ 0.
problem, and we need to get the primal-dual equilibrium V
variable, then the optimization functions (27)–(29) can be y
transformed as a set of linear variational inequalities, and to
find a primal-dual equilibrium vector X ∗ ∈ n and y∗ ∈ m, for Its time derivative along the neural network trajectory (33) can
example be obtained
 ∗   −   + 
X X ξ X ξ   ∗ T   
∈= | ≤ ≤ ⊂ Rn+m . dV X X I J
y∗ y −ω y +ω = − ∗ γ E+
dt y y −J T 0
(30)    
X I J X
× P −
Then, we have y −J T 0 y
  ∗ T  t   
X X λ 0 X(s)ds X
− ∗ + −
y y −ϒ y
  ∗  t       
I −J T X λ 0 X(s)ds X  X X I J X
· · ∗ + ≥ 0, ∀ ∈ ≤ −γ  − P −
J 0 y −ϒ y y y −J T 0 y
(31)  t 
λ 0 X(s)ds 2
+ 
where I ∈ Rn×n . Inspired by the work in [23], it is easy −ϒ 2
  ∗  
to know that linear variational inequality problem (31) is  X X 2
equivalent to the system of piecewise-linear equations, for − γ y − y∗
 ≤ 0.
2
example
  
X I −J T X From [23] and [30], the network state [X(t), y]T can be sta-
P − ·
y J 0 y ble and globally convergent to an equilibrium [X ∗ , y∗ ]T in view
 t  
λ 0 X(s)ds X of V̇ = 0 when [X ∗ , y∗ ]T = 0 and [X, y]T = [X ∗ , y∗ ]T . It can
+ − = 0 (32) be obtained that the first n elements of [X ∗ , y∗ ]T are the opti-
−ϒ y
mal solution to the optimization function (27)–(29). As for the
where P (.) : Rn+m →  is a projection operator. Then, exponential convergence, review V([X, y]T ), V̇([X, y]T ), and
from [30], the linear projection equation (32) can be computed the extra condition
by the following recurrent neural network as
     
Ẋ I J dV
X
  ∗ T
=γ E+ y
ẏ −J T 0 X X
     ≤ −γ − ∗
X I −J T X dt y y
× P −   
y J 0 y I J
 t    · σE +
λ 0 X(s)ds X −J T 0
+ −   ∗   
−ϒ y ·
X
− ∗
X
≤ −λV
X
(33) y y y
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

8 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

Fig. 10. sEMG signals in the relax status.


Fig. 13. Shoulder joint angle.

Fig. 11. sEMG signals in the grasp status.

Fig. 14. Elbow joint angle.

Fig. 12. sEMG signals in the extension status.

where λ = σ γ > 0 is convergence rate. Thus, V([X]T ) =


O(exp(−λ(t − t0 ))); ∀t ≥ t0 can be obtained, and hence Fig. 15. Wrist joint angle.
  t   t ∗ 
 
 0 X(t)dt − 0 X dt  2 = O(exp(−λ(t − t0 )/2))
 y y∗ 2

where ∀t ≥ t0 , which completes the proof of exponential sensors attached to the subject’s forearm, as shown in Fig. 8.
stability. Fig. 9 shows the three specific gesture: 1) relaxing; 2) grasp-
ing; and 3) full extension. During training, each posture is
held for 7 s with 5 s rest. The collected sEMG signals are
VI. E XPERIMENT V ERIFICATION shown in Figs. 10–12, respectively. After training, SVM clas-
To verify the effectiveness of the proposed approaches, the sifier would label three postures as follows: 1) 0 for relaxing;
developed 5 DOFs bionic arm is utilized in the experiments. 2) 1 for grasping; and 3) 2 for full extension. Actually, only
In the actual implementation, the sampling period is chosen as three fingers, for example, thumb, index finger, and middle
50 ms, and we choose three hand postures: 1) relaxing; 2) full finger, are involved in the experiment. There are four kinds
extension; and 3) grasping. First, the selected subject partici- of features extracted from four channels of sEMG signals, as
pates in the sEMG signal collection, there are the four sEMG shown in Table II. The four features were extracted from the
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

HU et al.: DEVELOPMENT OF SENSORY-MOTOR FUSION-BASED MANIPULATION AND GRASPING CONTROL 9

Fig. 16. Shoulder joint trajectory.


Fig. 19. Result of the classification.

Fig. 17. Elbow joint trajectory.


Fig. 20. Cloud of reachable points.

Fig. 18. End-effector trajectory.

Fig. 21. Manipulator trajectory.


segmented samples, and then used for training the feature pro-
jected matrices. Fig. 19 shows the accuracy and error rates of By detecting the five colored circular blobs of the par-
the three label classification. ticular mark, image coordinates of the feature point can be
The Bumblebee2 is placed on the top of the bionic shown robustly obtained by considering the color information and
in Fig. 1 such that the marker on the object can be tracked the known geometrical relationship among different blobs.
within the range of the camera perspective. The cameras Furthermore, through the use of image coordinates of the geo-
work at 20 frames/s and are calibrated before the experi- metrical centers of the colored circular blobs, the orientation
ments. The target object and the thumb of the hand have angle of the bionic arm on the image plane can also be calcu-
one circular features (red dots), respectively, shown in Fig. 7. lated in each control cycle. By considering the approximated
The trajectory and position of the objects are chosen care- distance of the camera as the depth of the feature point, we
fully to ensure the visibility of the objects for the cameras. can easily reconstruct the Cartesian coordinates of the feature
Fig. 20 shows the region labeled by the black lines of the point in the camera frame using its image coordinates and the
3-D point cloud. The one of the red marker is the target calibrated values of camera intrinsic parameters. In the exper-
object position, and the other is the initial position of the iments, the initial image coordinates and corresponding recon-
end-effector. structed Cartesian coordinates of the feature point are required
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

10 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

Fig. 22. Thumb trajectory. Fig. 25. Three fingers force.

2) pI = [−11.00, 11.00, 2.00] mm; and 3) pM =


[15.00, 11.00, 2.00] mm. The scaling parameter λf = 1000.
The transformation matrix G can be described as
⎡ ⎤
0 0 1 −0.011 −0.002 0
⎢1 0 0 0 0.002 0.011 ⎥
⎢ ⎥
⎢0 1 0 −0.002 0 0.002 ⎥
⎢ ⎥
⎢1 0 0 0 0.002 −0.011 ⎥
⎢ ⎥
GT = ⎢ ⎢0 0 1 0.011 0.011 0 ⎥ ⎥.
⎢ 0 −1 0 0.002 0 0.011 ⎥
⎢ ⎥
⎢1 0 0 0 0.002 −0.011 ⎥
⎢ ⎥
Fig. 23. Index and middle fingers trajectory. ⎣0 0 1 0.011 −0.015 0 ⎦
0 −1 0 0.002 0 −0.015

Force optimization solution can be solved with (24). The


mass of the target object is m = 0.51 kg. The friction
coefficient is chosen as μi = 0.8. We can calculate the corre-
sponding torque of each finger. By using the feedback of the
force sensor, we regulate the contact forces. Figs. 24 and 25
show the optimized contact force and control input motor
currents.

VII. C ONCLUSION
In this paper, a sensory-motor fusion-based manipulation
Fig. 24. Three fingers current. and grasping control for a robotic hand-eye system is devel-
oped, and it consists of three modules: 1) vision servoing;
2) sEMG-based movement cognition; and 3) hybrid force
to be stored for calculating the position and orientation of the and motion optimization for manipulation and grasping. The
end-effector. stereo camera is used to obtain the 3-D point cloud of a
The proposed motion optimization scheme (11)–(13) target object and provides the desired operational position.
with λ = 3 is used, where  = 1010 . The task The AdaBoost-SVM-based motion recognition is proposed to
duration time is 6 s, and the initial joint variables discriminate different movements based on sEMG of the multi-
q(0) = [0.1π, 0.1π, 0.1π, 0] in radians. The center of the fingered robotic hand. The operational space motion planning
bionic palm is the position of the end-effector. The length for bionic arm and force planning for multifingered robotic
of three links are: [26.50, 33.50, 15.00]T cm, n = 4, and hand can be both transformed as a convex optimization prob-
m = 3. Figs. 13–15 show the curves of three joints tracking lem, considering also various constraints. A neural dynamics
in joint space, respectively. The position tracking trajecto- optimization solution is proposed and is suitable to be imple-
ries in task space are presented in Figs. 16–18, respectively. mented online. Real-time experiments are conducted and the
From Figs. 13–18, it is shown that the real trajectories con- results demonstrate the responsiveness and flexibility of the
verges to the desired trajectories. Figs. 21–23 show the real proposed approach. In the future works, we shall consider
trajectory. The grasping model is defined by the position adaptive sliding mode control strategy for manipulation and
of three grasp points: 1) pT = [2.00, −11.00, 2.00]T mm; grasping of multifingered robotic hands [36]–[35].
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

HU et al.: DEVELOPMENT OF SENSORY-MOTOR FUSION-BASED MANIPULATION AND GRASPING CONTROL 11

R EFERENCES [22] Y. Xia, J. Wang, and L.-M. Fok, “Grasping-force optimization for mul-
tifingered robotic hands using a recurrent neural network,” IEEE Trans.
[1] R. de Souza, S. El-Khoury, J. Santos-Victor, and A. Billard, Robot. Autom., vol. 20, no. 3, pp. 549–554, Jun. 2004.
“Recognizing the grasp intention from human demonstration,” Robot. [23] Z. Li, S. Xiao, S. S. Ge, and H. Su, “Constrained multilegged robot sys-
Auton. Syst., vol. 74, pp. 108–121, Dec. 2015. tem modeling and fuzzy control with uncertain kinematics and dynamics
[2] C. Cipriani, M. Controzzi, and M. C. Carrozza, “The SmartHand tran- incorporating foot force optimization,” IEEE Trans. Syst., Man, Cybern.,
sradial prosthesis,” J. NeuroEng. Rehabil., vol. 8, no. 29, pp. 1–13, Syst., vol. 46, no. 1, pp. 1–15, Jan. 2016.
2011. [24] S. A. Dalley, T. E. Wiste, T. J. Withrow, and M. Goldfarb, “Design of
[3] X. Chen, D. Zhang, and X. Zhu, “Application of a self-enhancing a multifunctional anthropomorphic prosthetic hand with extrinsic actu-
classification method to electromyography pattern recognition for ation,” IEEE/ASME Trans. Mechatronics, vol. 14, no. 6, pp. 699–706,
multifunctional prosthesis control,” J. NeuroEng. Rehabil., vol. 10, no. 1, Dec. 2009.
2013, Art. no. 44. [25] R. Weir et al., “New multifunctional prosthetic arm and hand systems,”
[4] A. D. Roche, H. Rehbaum, D. Farina, and O. C. Aszmann, “Prosthetic in Proc. IEEE Int. Conf. EMBS, Lyon, France, 2007, pp. 4359–4360.
myoelectric control strategies: A clinical perspective,” Current Surg. [26] Z. Li, B. Wang, C. Yang, Q. Xie, and C.-Y. Su, “Boosting-based EMG
Rep., vol. 2, no. 3, pp. 1–11, Mar. 2014. patterns classification scheme for robustness enhancement,” IEEE J.
[5] C. Cipriani, F. Zaccone, S. Micera, and M. C. Carrozza, “On the Biomed. Health Informat., vol. 17, no. 3, pp. 545–552, May 2013.
shared control of an EMG-controlled prosthetic hand: Analysis of user– [27] H. Su and B. He, “A simple rectification method of stereo image pairs
prosthesis interaction,” IEEE Trans. Robot., vol. 24, no. 1, pp. 170–184, with calibrated cameras,” in Proc. 2nd Int. Conf. Inf. Eng. Comput.
Feb. 2008. Sci. (ICIECS), Wuhan, China, 2010, pp. 1–4.
[6] A. J. Young, L. H. Smith, E. J. Rouse, and L. J. Hargrove, “A com- [28] H. Hirschmuller, “Accurate and efficient stereo processing by semi-
parison of the real-time controllability of pattern recognition to conven- global matching and mutual information,” in Proc. IEEE Comput. Soc.
tional myoelectric control for discrete and simultaneous movements,” Conf. Comput. Vis. Pattern Recognit., vol. 2. San Diego, CA, USA,
J. NeuroEng. Rehabil., vol. 11, no. 1, 2014, Art. no. 5. Jun. 2005, pp. 807–814.
[7] M. Ortiz-Catalan, B. Håkansson, and R. Brånemark, “Real-time and [29] M. S. Bazaraa, H. D. Sherali, and C. M. Shetty, Nonlinear Programming
simultaneous control of artificial limbs based on pattern recognition Theory and Algorithms. New York, NY, USA: Wiley, 2013.
algorithms,” IEEE Trans. Neural Syst. Rehabil. Eng., vol. 22, no. 4, [30] Z. Zhang and Y. Zhang, “Acceleration-level cyclic-motion generation of
pp. 756–764, Jul. 2014. constrained redundant robots tracking different paths,” IEEE Trans. Syst.,
[8] J. Hahne et al., “Linear and nonlinear regression techniques for simulta- Man, Cybern. B, Cybern., vol. 42, no. 4, pp. 1257–1269, Aug. 2012.
neous and proportional myoelectric control,” IEEE Trans. Neural Syst. [31] Z. Li, C. Yang, C.-Y. Su, J. Deng, and W. Zhang, “Vision-based model
Rehabil. Eng., vol. 22, no. 2, pp. 269–279, Mar. 2014. predictive control for steering of a nonholonomic mobile robot,” IEEE
[9] E. Scheme, A. Fougner, Ø. Stavdahl, A. D. C. Chan, and K. Englehart, Trans. Control Syst. Technol., vol. 24, no. 2, pp. 553–564, Mar. 2016.
“Examining the adverse effects of limb position on pattern recognition [32] W. He, Y. Dong, and C. Sun, “Adaptive neural impedance control of
based myoelectric control,” in Proc. IEEE Eng. Med. Biol. Annu. Int. a robotic manipulator with input saturation,” IEEE Trans. Syst., Man,
Conf. (EMBC), Buenos Aires, Argentina, 2010, pp. 6337–6340. Cybern., Syst., vol. 46, no. 3, pp. 334–344, Mar. 2016.
[10] P. K. Artemiadis and K. J. Kyriakopoulos, “An EMG-based robot control [33] Z. Li, H. Xiao, C. Yang, and Y. Zhao, “Model predictive control of
scheme robust to time-varying EMG signal features,” IEEE Trans. Inf. nonholonomic chained systems using general projection neural networks
Technol. Biomed., vol. 14, no. 3, pp. 582–588, May 2010. optimization,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 45, no. 10,
[11] L. Hargrove, K. Englehart, and B. Hudgins, “The effect of electrode dis- pp. 1313–1321, Oct. 2015.
placements on pattern recognition based myoelectric control,” in Proc. [34] E. A. Al-Gallaf, “Neuro-kinematics based dexterous robotics hand
28th Annu. Int. Conf. IEEE Eng. Med. Biol. Soc., New York, NY, USA, force optimization,” J. Intell. Robot. Syst., vol. 50, no. 2, pp. 181–206,
2006, pp. 2203–2206. Oct. 2007.
[12] A. Gijsberts et al., “Stable myoelectric control of a hand prosthesis [35] H. Li, Z. Chen, L. Wu, H. K. Lam, and H. Du, “Event-triggered fault
using non-linear incremental learning,” Front. Neurorobot., vol. 8, no. 8, detection of nonlinear networked systems,” IEEE Trans. Cybern., to be
pp. 1–15, 2014. published, doi: 10.1109/TCYB.2016.2536750.
[36] H. Li, J. Wang, H.-K. Lam, Q. Zhou, and H. Du, “Adaptive sliding
[13] J. W. Sensinger, B. A. Lock, and T. A. Kuiken, “Adaptive pattern recog-
mode control for interval type-2 fuzzy systems,” IEEE Trans. Syst., Man,
nition of myoelectric signals: Exploration of conceptual framework and
Cybern., Syst., to be published, doi: 10.1109/TSMC.2016.2531676.
practical algorithms,” IEEE Trans. Neural Syst. Rehabil. Eng., vol. 17,
[37] H. Li, J. Wang, and P. Shi, “Output-feedback based sliding mode control
no. 3, pp. 270–278, Jun. 2009.
for fuzzy systems with actuator saturation,” IEEE Trans. Fuzzy Syst.,
[14] P. M. Pilarski et al., “Online human training of a myoelectric prosthesis to be published, doi: 10.1109/TFUZZ.2015.2513085.
controller via actor-critic reinforcement learning,” in Proc. IEEE Int.
Conf. Rehabil. Robot., Zürich, Switzerland, 2011, pp. 1–7.
[15] R. N. Khushaba, “Correlation analysis of electromyogram signals for
multiuser myoelectric interfaces,” IEEE Trans. Neural Syst. Rehabil.
Eng., vol. 22, no. 4, pp. 745–755, Jul. 2014.
[16] V. Kamat and S. Ganesan, “A robust hough transform tech-
nique for description of multiple line segments in an image,” in
Proc. Int. Conf. Image Process., vol. 1. Chicago, IL, USA, 1998,
pp. 216–220.
[17] J. Kofman, X. Wu, T. J. Luu, and S. Verma, “Teleoperation of a robot
manipulator using a vision-based human-robot interface,” IEEE Trans.
Ind. Electron., vol. 52, no. 5, pp. 1206–1219, Oct. 2005.
[18] C. Yang et al., “Visual servoing control of baxter robot arms with
obstacle avoidance using kinematic redundancy,” in Intelligent Robotics
and Applications (LNCS 9244). Cham, Switzerland: Springer, 2015,
pp. 568–580.
[19] H. Xiao, Z. Li, C. Yang, W. Yuan, and L. Wang, “ RGB-D sensor-based
visual target detection and tracking for an intelligent wheelchair robot Yingbai Hu received the B.S. degree in automation
in indoors environments,” Int. J. Control Autom. Syst., vol. 13, no. 3, from the Hubei University of Technology, Hubei,
pp. 521–529, Jun. 2015. China, in 2014. He is currently pursuing the master’s
[20] Z. Li et al., “sEMG-based joint force control for an upper-limb power- degree with the College of Automation Science and
assist exoskeleton robot,” IEEE J. Biomed. Health Inform., vol. 18, no. 3, Engineering, South China University of Technology,
pp. 1043–1050, May 2014. Guangzhou, China.
[21] Z. Li, B. Wang, C. Yang, Q. Xie, and C.-Y. Su, “Boosting- He has authored one paper in international confer-
based EMG patterns classification scheme for robustness enhance- ences. His current research interests include bionic
ment,” IEEE J. Biomed. Health Inform., vol. 17, no. 3, pp. 545–552, arms and prosthesis hands, and electromyography
May 2013. signals processing.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

12 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS

Zhijun Li (M’07–SM’09) received the Ph.D. Peijiang Yuan received the B.S. degree from
degree in mechatronics from Shanghai Jiao Tong Tsinghua University, Beijing, China, in 1997, and
University, Shanghai, China, in 2002. the Ph.D. degree in electronic and computer engi-
From 2003 to 2005, he was a Post-Doctoral neering from the University of Western Ontario,
Fellow with the Department of Mechanical London, ON, Canada, in 2005.
Engineering and Intelligent Systems, University He joined Beihang University, Beijing, in 2009,
of Electro-Communications, Tokyo, Japan. From where he is currently an Associate Professor with
2005 to 2006, he was a Research Fellow with the Intelligent Technology and Robotics Research
the Department of Electrical and Computer Center. He has authored over 50 papers in interna-
Engineering, National University of Singapore, tional journals and conferences. His research inter-
Singapore, and Nanyang Technological University, ests include aviation manufacturing robot, robot
Singapore. From 2007 to 2011, he was an Associate Professor with the localization and perception, intelligent control, and machine vision.
Department of Automation, Shanghai Jiao Tong University. In 2008, he Dr. Yuan has served for several international conferences as a Program
was a Visiting Scholar with Microsoft Research Asia, Beijing, China. Committee Member, the Program Committee Co-Chair of the 2010 IEEE
Since 2012, he has been a Professor with the College of Automation World Congress on Computational Intelligence, and a Committee Member of
Science and Engineering, South China university of Technology, Guangzhou, the 2011 IEEE Robotics and Automation and Intelligent Robots and Systems.
China. In 2015, he was a Visiting Professor with the Faculty Science and
Technology, University of Macau, Macau, China, and the Department of
Advanced Robotics, Italian Institute of Technology, Genoa, Italy. His current
research interests include service robotics, tele-operation systems, nonlinear
control, and neural network optimization. Chenguang Yang (M’10) received the B.Eng.
Dr. Li serves as an Editor-at-Large for the Journal of Intelligent and degree in measurement and control from
Robotic Systems, and an Associate Editor of the IEEE T RANSACTIONS Northwestern Polytechnical University, Xi’an,
ON N EURAL N ETWORKS AND L EARNING S YSTEMS and the IEEE China, in 2005, and the Ph.D. degree in control
T RANSACTIONS ON S YSTEMS , M AN , AND C YBERNETICS : S YSTEMS. engineering from the National University of
Singapore, Singapore, in 2010.
He is a Senior Lecturer with the Zienkiewicz
Centre for Computational Engineering, Swansea
University, Swansea, U.K. He received postdoctoral
Guanglin Li (M’01–SM’06) received the B.S. training with the Department of Bioengineering
and M.S. degrees in electrical engineering from at Imperial College London, London, U.K.
Shandong University, Jinan, China, in 1983 and His research interests include robotics, automation, and computational
1988, respectively, and the Ph.D. degree in intelligence.
biomedical engineering from Zhejiang University,
Hangzhou, China, in 1997.
From 2006 to 2009, he was a Senior Research
Scientist with the Neural Engineering Center for
Artificial Limbs, Rehabilitation Institute of Chicago,
Chicago, IL, USA, and a Research Assistant Rong Song received the B.S. degree in electri-
Professor with the Department of Physical Medicine cal engineering from Tsinghua University, Beijing,
and Rehabilitation, Northwestern University, Chicago. Since 2009, he has been China, in 1999, the M.S. degree in electronic engi-
with the Shenzhen Institutes of Advanced Technology, Chinese Academy neering from Shantou University, Shantou, China,
of Sciences, Shenzhen, China, where he is currently a Professor with in 2002, and the Ph.D. degree in biomedical engi-
the Research Centre for Neural Engineering, Institute of Biomedical and neering from Hong Kong Polytechnic University,
Health Engineering. His current research interests include neural rehabili- Hong Kong, in 2006.
tation engineering, neuroprosthesis control, biomedical signal analysis, and He is currently an Associate Professor with the
computational biomedical engineering. School of Engineering, Sun Yat-sen University,
Prof. Li is an Associate Editor of the IEEE J OURNAL OF B IOMEDICAL Guangzhou, China. His current research interests
AND H EALTH I NFORMATICS and serves as an International Advisory Board include musculoskeletal modeling, biomedical signal
Member of the Journal of Physiological Measurement. processing, human motion analysis, and robot-assisted stroke rehabilitation.

Das könnte Ihnen auch gefallen