Sie sind auf Seite 1von 5

International Conference on Computing, Communication and Automation (ICCCA2015)

Human Gait Based Gender Identification


System Using Hidden Markov Model and
Support Vector Machines

Deepjoy Das Alok Chakrabarty


Department of Computer Science & Engineering Department of Computer Science & Engineering
National Institute of Technology Meghalaya National Institute of Technology Meghalaya
Shillong, Meghalaya, India Shillong, Meghalaya, India
deepjoy2002@gmail.com mcscalok@gmail.com

Abstract—The paper presents an approach towards of human gait data doesn't require the subject to be in
human gender recognition system. The Silhouettes close proximity with the sensor as in gender identifi-
from Center for Biometrics and Security Research cation using facial expression, Electroencephalography
(CASIA) gait database are segmented in order to iden- (EEG) etc. It also possesses disadvantages like change
tify major body points and to generate corresponding in footwear can have varying gait patterns [33]. Gait
point-light display. The features such as two dimen-
sional coordinates of major body points and joint
pattern also differ in patients suffering from diabetes,
angles are extracted from the point-light display. The Parkinson's disease etc. [34] [35].
features are classified using Hidden Markov Model
(HMM) and Support Vector Machines (SVM). The study
yields a recognition rate of 69.18% and 76.79% with 100
II. P REVIOUS W ORK
subject data using HMM and SVM respectively. There The first study on gender recognition conducted in
has been a significant improvement in recognition year 1977, demonstrated that observers were able to
accuracy using joint angles as the features.
recognize gender with a performance of 63% accuracy,
Keywords—Gender Recognition/Identification, when light sources mounted on joints during waking
Point-light (PL) display, Hidden Markov Model concluded that PL points are the sufficient cues for the
(HMM), Support Vector Machine (SVM), Human Gait. study [15]. Similar study conducted by Barclay et al.
[4] in year 1978, concluded with numerous profound
I. I NTRODUCTION results - 1) At most two gait cycles is required to
Perceiving actions, emotion, intentions, gender & determine gender from biological motion, gait cycles
identity of others is an amazing skill every human pos- less than that will result in reduction of performance,
sesses and human visual system provides an accurate 2) Manipulation in speed can have a stronger effect
interpretation of it. However not much is known about in altered gait recognition, 3) Blurring the dots of PL
how this biological relevant information is encoded in walker can cause recognition accuracy to be dropped,
visual motion patterns and is used in perception. In 4) Inversion of PL points by 180 degrees would result
1973, G. Johansson [1] [2] devised the term known in decreased accuracy. The inversion result concluded
as PL animation, in which an observer is able to in an interesting result that in male the ratio width
recognize a person familiar to them, when the motion of shoulder and hip is positive, in female the ratio is
is portrayed by small number of markers positioned on negative. This concludes that detecting shoulder & hip
the head and the joints of the body. A large number of width is sufficient for gender identification [4].
studies related to human perception have used Johans- Yoo et al. [36] used stick view, which can be gener-
son's point light display [1] [2]. Point-light (PL) display ated by connecting the point light points, and achieved
contains sufficient information to perceive actions [3], a recognition rate of 96% for 100 subjects using a
gender [4] [5] [6] [7] [8], emotions [9] [10] [11] [12] [13] support vector based classifier. Davis et al. [37] used
[14] and to identify individuals [15] [16] [17] [18]. point light display with an interesting adaptive three-
Study of gait abnormalities is also used in diagnosis mode PCA to extract point-light display and got a
of diabetes [19], Hydrocephalus [20], Osteoarthritis recognition accuracy rate of 95.5% in a 40 subject
[21], Orthopedic impairments [22] [23] and in diagnosis database.
of mental ailments such as patients with schizophrenia
Pollick [38] conducted experiment to determine hu-
[24] [25], depression [26] [27], Alzheimer's disease [28]
man efficiency for recognition of individuals using
[29] [30] and Parkinson's disease [31] [32].
Point light display. The estimates of efficiency of hu-
Gait based gender identification system has a sig- man performance were compared to artificial neural
nificant advantage over other techniques as collection network performance. The study expressed efficiency

ISBN:978-1-4799-8890-7/15/$31.00 ©2015 IEEE 268


International Conference on Computing, Communication and Automation (ICCCA2015)

as the squared ratio of human sensitivity to neural B. Segmentation & Body Point Extraction
network sensitivity.
The area and centroid of grayscale image I(x, y) is
determined using the raw moment of order 0 and order
III. O UR A PPROACH 1 using the equation

The silhouettes of CASIA gait database B is used, Mij = xi y i I(x, y) (4)
from which we have used silhouette data for 50 males x y
and 50 females. In total 100 subjects are used for the  
study. We follow a set of pre-processing step in order to The area and centroid are M00 & M10 M01
respec-
M00 , M00
extract the required features for gender identification,
scale the features & finally model the features using tively.
Hidden Markov Model (HMM) and Support Vector The max point Imax and min point Imin in vertical
Machines (SVM) respectively for classification of new direction of image I(x, y) of size N × M is defined
features. as the first and last dot in the vertical direction, we
scan for max and min in the vertical direction, keeping
horizontal coordinate same as x0 .
A. Silhouette Noise Removal
Noise in the form of white dots popularly known as 
salt noise are removed from the silhouette using mor- do {y = y − 1}
Imax = (5)
phological opening operation with structuring element while {I(x0 , y − 1) = I(x0 , y)} 2  y  y0
(SE) of size 6 × 6.
The opening of a set A by SE B, denoted by A ◦ B is 
defined as [39], do {y = y + 1}
Imin =
while {I(x0 , y + 1) = I(x0 , y)} y0  y  M − 1
A ◦ B = ((A  B) ⊕ B) (1) (6)
The height of the person is simply computed using
Where, A  B is defined as erosion of A by B (A and absolute difference of the max and min points.
B are sets in Z2 ) [39],
H = |Imax (x, y) − Imin (x, y)| (7)
A  B = {z | (B)z ⊆ A} (2)
The image is segmented using the height. For a
body height H, an estimate of head is 0% of H, neck is
A ⊕ B is defined as dilution of A by B (A and B are 13% of H, hand is between 35-50% of H, waist is 47%
sets in Z2 ) [39], of H, knee is 71.5% of H and foot is between 93-98% of
H. (All points are distances from horizontal axis) [17]
A ⊕ B = {z | (B  )z ∩ A = φ} (3) [36].
The image is segmented further in order to search
and locate other body points. We now search in the
horizontal direction keeping the fixed range of 35-50%
of H in vertical direction, to find the front and rear
hand point (in terms of distance from vertical axis)
respectively. Similarly, the front and rear coordinate
points of the knee and foot are identified [17].
Two dimensional coordinates of the body points
identified are as follows: Head, Neck, Front Hand, Rear
Hand, Centroid, Front Knee, Rear Knee, Front Foot,
and Rear Foot.

C. Linear Interpolation to Estimate the Invisible Body


Points
The body points such as front hand, rear hand
and leg point disappear behind the body in each gait
cycle, or in a multi-person walking scenario one of the
subjects disappears behind the other. In such cases, the
body points are to be estimated by linear interpolation
for those frames. Let the body points of frame t are to
Fig. 1. Silhouettes of CASIA database. Subjects to the left side of be estimated by linear interpolation of estimated body
figure are males & subjects to the right side of figure are females. points of the frame t−1 and frame t+1. Let (x1 , y1 ) and
(x3 , y3 ) be the 2-dimensional coordinates of frame t − 1

269
International Conference on Computing, Communication and Automation (ICCCA2015)

Euclidean distance between head and neck point, hor-


izontal and vertical distances between each of the
following points are taken as features: distance of front
hand and centroid point, rear hand and centroid point,
front knee and centroid point, rear knee and centroid
point, front foot and front knee point, rear foot and
rear knee point.
The set of 9 angles between the major body joint
were computed: Angle between the head and the body
[few persons walk with their heads bent], front hand
and the body [measures the amount of front arm
swing], rear arm and the body [measures the amount
of rear arm swing], front thigh and vertical axis [mea-
sures the amount of front knee swing], rear thigh
and vertical axis [measures the amount of rear knee
swing], front leg shin and vertical axis, rear leg shin
and vertical axis, front leg foot and horizontal axis,
rear leg foot and horizontal axis.
Height, area, velocity (computed using displace-
Fig. 2. Segmented silhouettes of CASIA database.
ment of centroid) along with 13 Euclidean distances
& 9 major joint angles forms a 25-dimensional feature
vector for a single frame which uniquely distinguishes
and frame t + 1. The estimation of coordinates (x2 , y2 ) it from other frame(s) of any silhouette data.
of frame t is given by
We consider 50 walking subjects of each gender. As
x1 + x 3 we need fixed size feature vectors as an input for SVM,
x2 = (8)
2 we consider 300 random frames for silhouette data of
each subject. Therefore, the size of the total feature
(x2 − x1 )(y3 − y1 ) vector is 15000 × 25.
y2 = + y1 (9)
x3 − x1

E. Feature Scaling
The feature vectors are scaled in the range of [0, 1].

M (i, j) − P (i)
M  (i, j) = f or1  i  K, 1  j  25 (10)
Q(i) − P (i)

Where, M and M  represents unscaled and scaled


feature matrix respectively, each of size 15000 × 25.
P and Q represent minimum and maximum value in
linear vector of size 15000. The values in the feature
matrix M  is in range [0, 1]. K denotes the number of
frames.

IV. G ENDER R ECOGNITION U SING SVM


A. HMM Modeling
Fig. 3. Identified head, hand (left & right), centroid, knee (left &
right), leg (left & right) points from segmented silhouettes of CASIA A continuous HMM consisting of set of N states
database. Subjects to the left side of the figure are males & subjects and M distinct observations v1 v2 ......vM can be distin-
to the right side of figure are females. guished using state transition probabilities A = {aij },
where aij = P (qt+1 = sj |qt = si ) , 1  i, j  N, state
conditional probabilities bj (k) = P (Ot = vk |st = j),
D. Total Feature Vector
k = 1, 2, ......M , where Ot is the observation and st is
The relative distances between the points were used the active state at time t, and initial state distribution
as a feature: The set of 13 distances are computed: πi = P (q1 = i) [40] [41].

270
International Conference on Computing, Communication and Automation (ICCCA2015)

B. SVM Modeling Body points for SVM


classification HMM
(out of Kα Kβ Kγ Kδ
SVMs are called as discriminative classifiers which 25-dimensions)
try to draw boundary between different sets of data FS-1:
by constructing a N dimensional hyperplane. Let xi be includes 23.48 23.02 23.98 23.62 23.58
the input data for the ith example and di is the cor-
responding target output, the equation of the decision D (H , N )
a b c

FS-2:
surface is given by, includes FS1
and features
wT x + b = 0 (11) below (5 of 25)
29.19 30.01 30.20 29.62 30.18
For a given weight vector w and bias b, the separation D(CV d ,F HV e )
D(CH f ,F HH g )
between the hyperplane is defined by the above equa-
D(CV ,RHV h )
tion and the closest data point is called the margin of D(CH,RHH i )
separation. SVM tries to find the hyperplane for which FS-3:
the margin of separation is maximized. The data points includes FS2
(xi , di ) which lie closest to the optimal hyperplane and features
are called support vectors. For this, non-linear data below (9 of 25)
38.86 40.12 42.75 42.07 41.38
kernel functions are used which map a given set of D(CV ,F KV j )
input vectors into a higher dimensional space for easy D(CH,F KH k )
classification. Some of the kernel functions used are as D(CV ,RKV l )
follows [42]: D(CH,RKH m )
FS-4:
• linear: K(xi , xj ) = xTi xj includes FS3
and features
• polynomial: K(xi , xj ) = (γ xTi xj + r)d , γ > 0 below (13 of 25)
45.19 46.69 53.29 52.51 51.69
D(CV ,F LV n )
• radial basis function: K(xi , xj ) = exp(-γ ||xi − D(CH,F LH o )
xj ||2 ), γ > 0 D(CV ,RLV p )
D(CH,RLH q )
• sigmoid: K(xi , xj ) = tanh(γ xTi xj + r) FS-5:
includes FS4
and features
Here, γ, r and d are kernel parameters. below (22 of 25) 61.71 63.92 71.75 71.76 69.58

We have used HTK Toolkit [43] and LIBSVM 9 Major


Toolkit [44] for Hidden Markov Model (HMM) and Sup- Joint Angles
port Vector Machine (SVM) modeling respectively. We FS-6:
divide the whole set of feature vector of size 15000 × 25 includes FS5
and features
into three parts as 60%, 20%, and 20% of the whole below (25 of 25) 69.18 69.15 76.79 75.74 72.98
set to obtain the training, cross-validation and testing
sets. Height
Area
Velocity
D a (x, y): Denotes Euclidean distance between element x and element y
H b : Denotes 2-dimensional coordinate of head, N c : Denotes 2-dimensional coordinate of neck,
V. C ONCLUSION CV d : Denotes vertical coordinate of centroid, CH e : Denotes horizontal coordinate of centroid,
F HV f : Denotes vertical coordinate of front hand, F HH g : Denotes horizontal coordinate of front hand,
RHV h : Denotes vertical coordinates of right hand, RHH i : Denotes horizontal coordinate of right hand,
We conclude that polynomial kernel of SVM yields F KV j : Denote vertical coordinate of front knee, F KH k : Denote horizontal coordinate of front knee,
RKV l : Denote vertical coordinate of right knee, RKH m : Denotes horizontal coordinate of right knee,
the highest recognition rate of 76.79% for our gender F LV n : Denotes vertical coordinate of front leg, F LH o : Denotes horizontal coordinates of front leg,
recognition system using 100 subject data. Almost all RLV p : Denotes vertical coordinate of right leg, RLH q : Denotes horizontal coordinate of right leg,
F S: Denotes feature set, K α : Denotes Linear Kernel, K β : Denotes Polynomial Kernel,
the kernels of SVM outperforms HMM at each stage K γ : Denotes RBF Kernel, K δ : Denotes Sigmoid Kernel.

of inclusion of new features. We further rank the TABLE I. R ECOGNITION A CCURACY OF HMM & SVM USING
SET OF DIFFERENT FEATURES.
RBF, Sigmoid, Linear kernels as second, third, fourth
performing SVM kernels respectively and HMM as
the least performing. It is also to be noted that the
inclusion of 9 major joint angles results in a drastic R EFERENCES
improvement in gender recognition accuracy.
[1] G. Johansson, “Visual perception of biological motion and a
model for its analysis,” Perception & psychophysics, vol. 14,
no. 2, pp. 201–211, 1973.
A CKNOWLEDGMENT [2] ——, “Visual motion perception.” Scientific American, vol. 232,
no. 6, pp. 76–88, 1975.
We thank Center for Biometric and Security Re- [3] W. H. Dittrich, “Action categories and the perception of biolog-
search (CBSR) for providing the open source CASIA ical motion,” perception-london-, vol. 22, pp. 15–15, 1993.
Gait Database with abundance of Gait Data for the [4] C. D. Barclay, J. E. Cutting, and L. T. Kozlowski, “Temporal
experiments. and spatial factors in gait perception that influence gender

271
International Conference on Computing, Communication and Automation (ICCCA2015)

recognition,” Perception & psychophysics, vol. 23, no. 2, pp. bidity of alcoholism,” American Journal of Psychiatry, vol. 161,
145–152, 1978. no. 4, pp. 751–755, 2004.
[5] H. Hill and A. Johnston, “Categorizing sex and identity from [25] A. Putzhammer, B. Heindl, K. Broll, L. Pfeiff, M. Perfahl,
the biological motion of faces,” Current Biology, vol. 11, no. 11, and G. Hajak, “Spatial and temporal parameters of gait dis-
pp. 880–885, 2001. turbances in schizophrenic patients,” Schizophrenia research,
[6] L. T. Kozlowski and J. E. Cutting, “Recognizing the sex of vol. 69, no. 2, pp. 159–166, 2004.
a walker from a dynamic point-light display,” Perception & [26] M. R. Lemke, T. Wendorff, B. Mieth, K. Buhl, and M. Linne-
Psychophysics, vol. 21, no. 6, pp. 575–580, 1977. mann, “Spatiotemporal gait patterns during over ground loco-
[7] G. Mather and L. Murdoch, “Gender discrimination in biolog- motion in major depression compared with healthy controls,”
ical motion displays based on dynamic cues,” Proceedings of Journal of psychiatric research, vol. 34, no. 4, pp. 277–283,
the Royal Society of London. Series B: Biological Sciences, vol. 2000.
258, no. 1353, pp. 273–279, 1994. [27] J. Michalak, N. F. Troje, J. Fischer, P. Vollmar, T. Heidenre-
[8] S. Runeson, “Perception of biological motion: The ksd-principle ich, and D. Schulte, “Embodiment of sadness and depression-
and the implications of a distal versus proximal approach,” gait patterns associated with dysphoric mood,” Psychosomatic
Perceiving events and objects, pp. 383–405, 1994. medicine, vol. 71, no. 5, pp. 580–587, 2009.
[9] D. Das, “An approach to emotion identification using human [28] J. Verghese, R. B. Lipton, C. B. Hall, G. Kuslansky, M. J. Katz,
gait,” in Proceedings of Fourth International Conference on Soft and H. Buschke, “Abnormality of gait as a predictor of non-
Computing for Problem Solving. Springer, 2015, pp. 161–171. alzheimer’s dementia,” New England Journal of Medicine, vol.
[10] Y. Ma, H. M. Paterson, and F. E. Pollick, “A motion capture 347, no. 22, pp. 1761–1768, 2002.
library for the study of identity, gender, and emotion perception [29] P. L. Sheridan, J. Solomont, N. Kowall, and J. M. Hausdorff,
from biological motion,” Behavior research methods, vol. 38, “Influence of executive function on locomotor function: divided
no. 1, pp. 134–141, 2006. attention increases gait variability in alzheimer’s disease,”
[11] A. P. Atkinson, W. H. Dittrich, A. J. Gemmell, A. W. Young Journal of the American Geriatrics Society, vol. 51, no. 11, pp.
et al., “Emotion perception from dynamic and static body 1633–1637, 2003.
expressions in point-light and full-light displays,” Perception- [30] S. O’keeffe, H. Kazeem, R. Philpott, J. Playfer, M. Gosney, and
London, vol. 33, no. 6, pp. 717–746, 2004. M. Lye, “Gait disturbance in alzheimer’s disease: a clinical
[12] W. H. Dittrich, T. Troscianko, S. E. Lea, and D. Morgan, study,” Age and Ageing, vol. 25, no. 4, pp. 313–316, 1996.
“Perception of emotion from dynamic point-light displays rep- [31] M. E. Morris, R. Iansek, T. A. Matyas, and J. J. Summers,
resented in dance,” Perception-London, vol. 25, no. 6, pp. 727– “The pathogenesis of gait hypokinesia in parkinson’s disease,”
738, 1996. Brain, vol. 117, no. 5, pp. 1169–1181, 1994.
[13] A. S. Heberlein, R. Adolphs, D. Tranel, and H. Damasio, “Cor- [32] B. R. Bloem, J. M. Hausdorff, J. E. Visser, and N. Giladi,
tical regions for judgments of emotions and personality traits “Falls and freezing of gait in parkinson’s disease: a review
from point-light walkers,” Journal of Cognitive Neuroscience, of two interconnected, episodic phenomena,” Movement Disor-
vol. 16, no. 7, pp. 1143–1158, 2004. ders, vol. 19, no. 8, pp. 871–884, 2004.
[14] R. D. Walk and C. P. Homan, “Emotion and dance in dynamic
[33] M. A. Nurse, M. Hulliger, J. M. Wakeling, B. M. Nigg, and
light displays,” Bulletin of the Psychonomic Society, vol. 22,
D. J. Stefanyshyn, “Changing the texture of footwear can alter
no. 5, pp. 437–440, 1984.
gait patterns,” Journal of electromyography and kinesiology,
[15] J. E. Cutting and L. T. Kozlowski, “Recognizing friends by their vol. 15, no. 5, pp. 496–506, 2005.
walk: Gait perception without familiarity cues,” Bulletin of the
psychonomic society, vol. 9, no. 5, pp. 353–356, 1977. [34] L. Allet, S. Armand, R. A. de Bie, Z. Pataky, K. Aminian, F. R.
Herrmann, and E. D. de Bruin, “Gait alterations of diabetic
[16] H. Hill and F. E. Pollick, “Exaggerating temporal differences patients while walking on different surfaces,” Gait & posture,
enhances recognition of individuals from point light displays,” vol. 29, no. 3, pp. 488–493, 2009.
Psychological Science, vol. 11, no. 3, pp. 223–228, 2000.
[35] A. H. Snijders, I. Leunissen, M. Bakker, S. Overeem, R. C.
[17] D. Das and S. Saharia, “Human gait analysis and recognition Helmich, B. R. Bloem, and I. Toni, “Gait-related cerebral
using support vector machines.” International Journal of Com- alterations in patients with parkinsons disease with freezing
puter Science & Information Technology, vol. 6, no. 5, 2014. of gait,” Brain, vol. 134, no. 1, pp. 59–72, 2011.
[18] D. Das, “Human gait classification using combined hmm & svm
[36] J.-H. Yoo, D. Hwang, and M. S. Nixon, “Gender classification
hybrid classifier,” in Electronic Design, Computer Networks &
in human gait using support vector machine,” in Advanced
Automated Verification (EDCAV), 2015 International Confer-
concepts for intelligent vision systems. Springer, 2005, pp.
ence on. IEEE, 2015, pp. 169–174.
138–145.
[19] A. Bellmeyer and J. Strasser, “Effect of stride frequency on
plantar loading in type 2 diabetes.” [37] J. W. Davis and H. Gao, “Gender recognition from walk-
ing movements using adaptive three-mode pca,” in Computer
[20] M. A. Williams, G. Thomas, B. de Lateur, H. Imteyaz, J. G. Vision and Pattern Recognition Workshop, 2004. CVPRW’04.
Rose, W. S. Shore, S. Kharkar, and D. Rigamonti, “Objective Conference on. IEEE, 2004, pp. 9–9.
assessment of gait in normal-pressure hydrocephalus,” Amer-
ican Journal of Physical Medicine & Rehabilitation, vol. 87, [38] F. E. Pollick, V. Lestou, J. Ryu, and S.-B. Cho, “Estimating
no. 1, pp. 39–45, 2008. the efficiency of recognizing gender and affect from biological
motion,” Vision research, vol. 42, no. 20, pp. 2345–2355, 2002.
[21] M. Peterson, P. Kovar-Toledano, J. Otis, J. Allegrante,
C. Mackenzie, B. Gutin, and M. Kroll, “Effect of a walking [39] R. C. Gonzalez and R. E. Woods, Digital image processing, 2nd,
program on gait characteristics in patients with osteoarthri- 2002, vol. 2.
tis,” Arthritis & Rheumatism, vol. 6, no. 1, pp. 11–16, 1993. [40] L. Rabiner, “A tutorial on hidden markov models and selected
[22] J. J. Brunnekreef, C. J. van Uden, S. van Moorsel, and J. G. applications in speech recognition,” Proceedings of the IEEE,
Kooloos, “Reliability of videotaped observational gait analysis vol. 77, no. 2, pp. 257–286, 1989.
in patients with orthopedic impairments,” BMC musculoskele- [41] D. Das, R. Devi, S. Prasanna, S. Ghosh, and K. Naik, “Perfor-
tal disorders, vol. 6, no. 1, p. 17, 2005. mance comparison of online handwriting recognition system
[23] K. B. Tracy, E. Montague, R. Gabriel, and B. Kent, “Computer- for assamese language based on hmm and svm modelling.”
assisted diagnosis of orthopedic gait disorders.” Physical ther- International Journal of Computer Science & Information
apy, vol. 59, no. 3, pp. 268–277, 1979. Technology, vol. 6, no. 5, 2014.
[24] E. V. Sullivan, M. J. Rosenbloom, and A. Pfefferbaum, “Balance [42] C.-W. Hsu, C.-C. Chang, C.-J. Lin et al., “A practical guide to
and gait deficits in schizophrenia compounded by the comor- support vector classification,” 2003.

272

Das könnte Ihnen auch gefallen