ArticlePDF Available

Gait Recognition System Tailored for Arab Costume of The Gulf Region

Authors:

Abstract

Existing work on gait recognition is focused on casual (western) customs hence not suitable for the gulf region where long gowns are used for both men and women. This paper proposes a gait recognition solution that is suitable for both gulf customs and casual customs. The solution is based on computing an adaptive image prediction between consecutive images. The resultant predictions are then accumulated into one image and transformed using either discrete cosine transformation (DCT) or Radon transformation. The feature vectors of the gait are computed from such transformed images. Feature modeling based on polynomial networks follows. The proposed solution is tested on a dataset with around 100 participants with mixed genders and mixed customs. The proposed system yields an impressive classification rates approaching 100% accuracy.
Gait Recognition System Tailored for Arab Costume of The Gulf Region
Tamer Shanableh
Department of computer
science and Engineering
American University of
Sharjah
tshanableh@aus.edu
Khaled Assaleh
Department of Electrical
Engineering
American University of
Sharjah
kassaleh@aus.edu
Layla Al – Hajjaj
Computer Science
program
American university of
Sharjah
g00021907@aus.edu
AbdulWahab Kabani
Computer Science
program
American university of
Sharjah
b00020950@aus.edu
Abstract –
Existing work on gait recognition is focused on casual
(western) customs hence not suitable for the Gulf region where long gowns
are used for both men and women. This paper proposes a gait recognition
solution that is suitable for both Gulf customs and casual customs. The
solution is based on computing an adaptive image prediction between
consecutive images. The resultant predictions are then accumulated into one
image and transformed using either Discrete Cosine Transformation
(DCT) or Radon transformation. The feature vectors of the gait are
computed from such transformed images. Feature modeling based on
polynomial networks follows. The proposed solution is tested on a dataset
with around 100 participants with mixed genders and mixed customs. The
proposed system yields an impressive classification rates approaching 100%
accuracy.
Keywords –
Human identification; computer vision; motion analysis;
gait biometric
1. INTRODUCTION
In biometrics, people are identified based on their
characteristics such as voice, iris, fingerprint, hand geometry
and face. It has been reported that such identification can also
be based on the way that a human walks [1]. Such a biometric is
referred to as Gait. Basically video cameras are used to acquire
video sequences of individuals and recognize them based on the
way they walk. Gait recognition has a number of attractive
characteristics when compared to existing biometrics [2]. For
example it does not require a physical contact as required by
fingerprint or hand recognition. It also does not require high
image resolution or special image acquisition conditions as
required by face recognition for instance. Lastly it is non-
intrusive and can recognize people at a distance without their
knowledge or direct involvement.
In 2005, a research group from the University of South Florida
issued a human gait recognition challenge [3]. The group
compiled a dataset of video sequences with different covariates
such as camera viewing angle, walking surface type, carrying
conditions where a person can be carrying a briefcase for
example, shoe type where walking in heals for instance will
affect the gait and the video capturing time. In the latter most
video sequences were acquired in a second round after six
months of the first shooting. The dataset contains data for
experiments of increasing difficulty levels. A total of 1,870
video sequences are acquired with a total of 122 participants.
The dataset is made available and it is used for benchmarking
new solutions in gait recognition.
However, this dataset and subsequently all the previously
proposed solutions are based on western style of dress code.
Such solutions are likely to fail when applied to Gulf style dress
code including white/black robes, head gears/scarves and veils.
This is so because of the nature of the feature extraction
methods that exploit the gait cycle which depends on the
movements of the legs. .
This paper proposes an efficient feature extraction and
classification scheme for gait recognition for the Arab costume
in the Gulf region. The proposed scheme is also shown to work
for western costume as well.
In general gait recognition based on video sequences is divided
into a number of steps:
1. Segmentation: this step entails identifying the pixel locations
belonging to the subject to be identified. The segmented images
are binarized resulting in what is known as “silhouette frames”.
One approach for this segmentation is though background
modeling and separation. For instance, in [3] it was proposed to
extract bounding boxes of the subjects and then compute the
mean vector and covariance matrix of background pixels. The
pixels of the bounding box containing the subject are then
classified into either foreground or background using
Mahalanobis distances from the background model. The
distances are then classified into foreground or background
based on their likelihoods which are estimated using an
Expectation Minimization (EM) procedure.
Variants of this segmentation algorithm are also reported in the
literature. For instance, in [5] the pixels’ Mahalanobis distances
from the background model were thresholded into either
foreground or background without the need for computing
likelihoods through the EM algorithm. Other approaches
include extracting principal components of silhouette boundary
vector variations [6] or Fourier descriptors [7].
2. Feature extraction: This step can be preceded by what is
known as gait cycle estimation which is the set of images
starting from the right heel touching the ground all the way to
where it touches the ground again. This information can be
used to segment the sequences into cycles and then align them
using various techniques such as Population Hidden Markov
Models (PHMMs) [4]. Features are then based on averaged
subsequences [8]. Other feature extraction techniques are
reported for characterizing gait dynamics such as stride, stride
speed, length and cadence [9]. Others used static body
information such as ratios of various body parts [9]. Feature
vectors composed of the amplitude of the spectrum of key
silhouette frames are also reported in [2]. More recently
dynamic features from averaged silhouette cycles are extracted
by Gabor-based discriminative common vectors (DCV) analysis
[10]. Likewise [11] proposed the use of Kernel-based Principal
Component Analysis (KPCA) to extract gait features. In other
approaches, the human body components are studied separately
and feature vectors are extracted accordingly [12].
3. Feature modeling and similarity measures: Here the extracted
features are compared against the stored entries in the dataset.
Reported measures include computing the Euclidean distances
in the Linear Discriminate Analysis (LDA) space [4, 13],
symmetric group theoretic distances [14], normalized Euclidean
distance between the projection centroids of two gait sequences
[15]. Dynamics of the gait sequences can also be modeled by
hidden Markov models (HMMs) as reported in [16].
The existing work on gait recognition is however based on
identifying people in western or casual costumes namely; pants
and shorts. Such solutions are not suitable for identifying
individuals in the Gulf region of the Middle East. The local
dress code in the Gulf region for males includes robes
and head gears. Likewise the dress code for females includes
robes and head scarves or face veils. Examples of such customs
are shown in Figure 1. It is clear that the gap between the legs
of the individuals are concealed hence all the techniques based
on gait cycles do not apply. Note that this problem can also be
present if the individuals are dressed in long skirts as well;
hence the problem is not specific for the Gulf customs. We
propose a solution that applies to both casual and Gulf customs
based on accumulating the adaptive prediction errors of
consecutive images are shall be explained in Section 3.
The rest of this paper is organized as follows. The compiled
dataset and data acquisition procedure is described in Section 2.
Feature extraction and motion representation for causal and
Gulf customs are presented in Section 3. The classification
problem is then formalized using polynomial networks in
Section 4. Experimental results are presented and discussed in 5
prior to arriving to the final conclusion in section 6.
2. DATASET DESCRIPTION
Although the purpose of this research is to devise a method for
gait recognition for individuals in Gulf costumes nonetheless,
we need to verify that the proposed solution is also applicable
to recognizing individuals in casual costumes (mainly with pants
or shorts). As such the same system can be deployed for
identifying individuals with mixed costumes.
Similar to the setup reported in [3] the camera was positioned
10 meters away from the walking subjects. However one digital
camera was used with one view only. The video capturing took
place in the rotunda of one of our lecture building using one
digital camera. An example of participants with different
costumes is shown in Figure 1.
(a)
(b)
(c)
(d)
Fig. 1. Example participants with different costumes. (a) Female
with Gulf costume (b) Male with Gulf costume (c) Female with
casual costume (d) Male with casual costume.
A total number of 103 subjects participated in the data
collection. All participants are undergraduate students of the
same age group between 18 and 22 year old. Out of the 103
subjects, 53 participated with Gulf costumes (33 females and 20
males). Another 50 subjects participated in casual costumes (11
females and 39 males).
Each participant was asked to walk naturally across the rotunda
back and forth a total number of 8 times. Out of which 4
instances are captured with a walk from right to left and 4 in the
other direction.
3. FEATURE EXTRACTION
The existing literature on Gait recognition heavily depends on
the extraction of gait cycles. Such a cycle can be defined as the
sequence of images from which the right heel of an individual
touched the ground all the way until it touches the ground
again. The extraction of gait cycles depend on the observing the
g
a
c
e
s
a
b
e
a
p
W
a
c
b
e
t
h
s
u
c
o
o
b
t
h
p
r
b
e
b
i
O
m
I
n
u
s
p
r
is
p
r
d
i
i
n
p
e
t
h
i
n
P
r
i
n
c
o
t
h
d
i
s
e
c
a
N
a
i
n
m
i
m
p
r
b
a
o
f
t
h
Th
r
e
fi
g
n
o
a
p between the
e
rtain position
a
me position.
U
e
tween the le
g
p
proach to feat
u
W
e propose t
o
c
cumulate it in
t
e
extracted fro
m
h
at in the da
u
bjects are w
a
o
ntains variou
s
b
ject extractio
n
h
e other hand,
r
eprocessing
s
e
longing to t
h
i
narized result
i
O
ne approach
f
m
odeling and se
p
n
this paper w
e
s
ed in digital
v
r
ediction error
subtracted fr
o
r
ediction erro
r
i
fferences tha
t
n
dividual. The
e
rcentile of no
n
h
resholded pre
d
n
to one imag
e
r
edictions (A
P
n
dividual’s mot
i
o
nsecutive ima
g
h
e positive d
i
i
fferences. Ea
c
e
parately. In t
h
a
n be referred t
N
ote that the co
relative motio
n
n
the AP ima
g
m
otion we pro
p
m
ages in comp
u
r
ediction in th
i
a
ckward predic
f
forward or b
a
h
e Sum of Abs
o
h
e prediction
s
e
sult of imple
m
g
ure shows th
a
o
w minimized
a
two legs in a
v
of the legs an
d
U
nfortunately,
g
s is not at
u
re extraction
s
o
extract the
t
o one or two
i
m
such image
s
t
aset descripti
o
a
lking in fron
s
stationary o
b
n
and segment
a
in the absenc
e
s
hall entail i
d
e subject. Th
e
ng in what is
f
or this segm
e
p
aration as des
c
e
base our feat
u
v
ideo coding
w
between succe
o
m its immedia
r
can be th
r
t
did not re
s
threshold can
n
-zero pixels o
f
d
iction error i
m
e
which we
P
) image. Fo
r
i
on, the forwa
r
g
es can be rep
r
i
fferences an
d
c
h prediction
e
h
is case we en
d
o as positive A
v
ered and unc
o
n
to the indivi
d
g
es.
T
o mini
m
p
ose to use th
e
u
ting the predi
c
i
s case is refer
r
tion respective
l
a
ckward predic
o
lute Differenc
s
ource that mi
n
m
enting this tec
h
a
t the appearan
c
a
s desired.
(
v
ideo sequence
d
ends after a
f
with the Gulf
all apparent
h
s
hall be sought
.
motion of a
n
i
mages. Featur
e
s
that describe
o
n it was m
e
t of a static
b
jects. Hence
t
a
tion is not ne
e
e
of a stationa
r
d
entifying the
e
segmented i
m
known as “s
i
e
ntation is thr
o
c
ribed above [
3
u
re extraction
o
w
here we co
m
s
sive images.
T
te previous im
a
r
esholded to
s
ult from the
be set to th
e
f
the predictio
n
m
ages can the
n
refer to as
t
r
better repre
r
d prediction
e
r
esented using
2
d
the other
f
e
rror image is
d
up with two
P
and negative
o
vered backgro
d
ual and thus r
e
m
ize the appe
a
e
previous im
a
c
tion error for
a
r
ed to as forw
a
l
y.
T
he decisio
n
tion can be ba
s
e
s (SAD) of th
e
n
imizes the SA
D
h
nique is show
n
c
e of the back
g
(
a)
that starts fro
m
f
ull cycle with
costume, the
g
h
ence a diffe
r
.
n
individual
a
e
vectors can t
h
the motion.
N
e
ntioned that
background
t
t
he preprocess
e
ded this case.
r
y background
pixel locati
o
m
ages are usu
i
lhouette fram
e
o
ugh backgro
u
3
]..
o
n the techni
q
m
pute the for
w
T
hat is, each
i
m
a
ge. The resul
t
filter out im
motion of
e
50
th
or the
7
n
error image.
T
n
be accumul
a
t
he Accumul
a
sentation of
e
rror between
t
2
images. One
f
or the nega
t
then threshol
d
AP images w
h
AP images.
und will appea
r
e
presented as s
u
a
rance of rela
t
a
ges or the fu
t
a
given image.
T
a
rd prediction
a
n
between the
s
ed on compu
t
e
prediction er
r
D
is selected.
T
n
in Figure 2.
T
g
round objects
m
a
the
g
ap
r
ent
a
nd
h
en
N
ote
the
t
hat
of
On
the
o
ns
ally
e
s”.
u
nd
q
ues
w
ard
age
t
ant
age
the
7
5
th
T
he
a
ted
a
ted
the
t
wo
for
t
ive
d
ed
h
ich
r
as
u
ch
t
ive
t
ure
T
he
a
nd
use
t
ing
r
or.
T
he
T
he
are
Fi
g
Once t
h
spatial
d
languag
e
based
T
ransf
o
coeffici
e
transfo
r
energy
i
i
maged.
such l
o
quantiz
a
content
.
propos
e
coeffici
e
scannin
g
progres
s
scannin
g
coeffici
e
can be
s
T
he p
r
scannin
g
Note t
h
positive
process
A
s me
n
extracti
o
A
P im
a
dimens
i
the dire
done o
n
smooth
one di
m
i
deal lo
w
projecti
o
coeffici
e
Polyno
m
which
n
higher
sequen
c
main s
t
feature
i
mprov
i
expand
e
comput
i
g
. 2. AP image
s
forw
a
(a) Negative
A
h
e AP images
a
d
omain featur
e
e recognition
a
on either t
h
o
rm (DCT)
c
e
nts.
A
n
r
mation is its
i
s concentrate
d
This fact is
u
o
w frequency
a
tion step size
. Therefore i
n
e
to represent
e
nts. There co
e
g
manner st
a
sing inwards
t
g
process c
a
e
nts. This nu
m
s
elected empiri
c
r
ocess of D
C
g
is also know
n
h
at the zonal co
e
AP images. T
h
are then interl
e
n
tioned previ
o
o
n is based o
n
a
ges are projec
i
onal curve tha
t
e
ction of the p
r
n
either the
h
and reduce th
m
ensional DC
T
w
pass filtering
o
n can be
r
e
nts.
4.
m
ial network
p
n
onlinearly ex
p
dimensionalit
y
c
e. Training o
f
t
ages. The fir
s
vectors via
p
i
ng the separ
e
d feature v
e
t
ing the weigh
t
(b)
s
of a motion s
e
a
rd/backward
p
A
P image (b)
P
a
re computed,
e
s. Following
t
a
s reported in
h
e two dim
e
c
oefficients o
r
important
p
energy compa
d
in the top left
u
tilized in ima
g
coefficients a
r
in compariso
n
n
terms of sp
a
our AP imag
e
e
fficients can
b
a
rting from
t
towards the
b
a
n select a
m
ber is known
c
ally.
C
T transforma
t
n
as Zonal cod
i
o
ding is applied
h
e resultant ve
c
e
aved to gener
a
o
usly, the sec
o
n
Radon trans
f
c
ted at a given
t reflects the i
n
r
ojection angle.
h
orizontal or t
h
h
e size of the c
o
T
transformatio
with a given f
r
r
epresented u
CLASSIFIC
A
p
rovides a par
p
ands a seque
n
y
and maps
t
f
a polynomia
l
s
t stage involv
p
olynomial ex
p
r
ation of the
e
ctor space.
T
t
s of the poly
n
e
quence with a
d
p
rediction.
P
ositive AP im
a
the next step
t
he authors’
w
[
17
]
, these fea
t
e
nsional Disc
r
r
the Rado
n
p
roperty of
ction. Most o
f
t
corner of the
g
e and video c
o
r
e quantized
w
n
with the hi
g
a
tial feature e
x
e
using the t
o
b
e selected usi
n
t
he top left
b
ottom right
c
predefined
n
as the DCT
c
t
ion followed
i
ng.
to both negati
v
c
tors of the zo
n
a
te the final fe
a
o
nd approach
f
ormation. Es
s
angle; the re
s
n
tegral of pixel
T
ypically the
p
h
e vertical im
a
o
mputed imag
e
n can be used
r
equency cuto
ff
sing few lo
w
A
TION
r
ameterized no
n
n
ce of feature
v
t
hem to a ta
l
network con
s
v
es expanding
p
ansion with
different cla
s
T
he second s
n
omial networ
k
d
aptive
a
ge.
is to extract
w
ork on sing
t
ures can be
r
ete Cosine
n
transform
the DCT
f the image
transformed
o
ding where
w
ith a finer
g
h frequency
x
traction we
o
p left DCT
n
g a zig-zag
corner and
c
orner. The
n
umber of
c
utoff which
by zigzag-
v
e and
n
al coding
a
ture vector.
to feature
s
entially, the
s
ult is a one
lines across
p
rojection is
a
ge axis. To
e
projection,
followed by
ff
. Hence the
w
frequency
n
linear map
v
ectors to a
a
rget output
s
ists of two
the training
the aim of
s
ses in the
tage entails
k
applied to
the expanded feature vectors. Polynomial networks have been
used successfully in biomedical signal separation [18].
In a polynomial networks setting, the gait recognition problem
can be formulated as follows. The response variables which
represent the M individuals (where each individual can be
referred to a class in this case) of the training dataset, are
denoted by an M number of q vectors, i.e. ={| =
1,2,,} . For a given class of features vectors, say class i, the
corresponding q vector will contain binary values with ‘1s’
indicating individuals belonging to class i and ‘0s’ for the rest
of the individuals or participants.
The feature vector of the ith individual at repetition j of class m
is composed of l feature variables and is denoted by
,=
[
,(0)
,(1)…
,()]. Consequently, the feature
vectors in the training set is denoted by the matrix X where:

,(0)
,(1)⋯
,()
,(0)
,(1)…
,()
⋮⋮
,(0)
,(1)…
,()
(1)
We wish to perform a nonlinear mapping between the feature
vector matrix X and the response variables ={| =
1,2,,}. In polynomial networks, the dimensionality of the
feature vectors in matrix X is first expanded into an rth order.
The dimensionality expansion can be achieved by a reduced
multivariate polynomial expansion as proposed in [16]. The
expansion of X into the rth order is denoted by the matrix P
Rn x k where k is the dimensionality of the expanded feature
vector which is defined as [16]:
k = 1+r + l(2r-1) (2)
The mapping between P and is then achieved by using least-
squared error objective criterion:
 =avg
min −
(3)
Where . denotes the l2 norm. Minimizing the objective
function results in:
 =(
) (4)
Note that model weights are computed using a non-iterative
least squares method which is a clear advantage when it comes
to computational complexity.
Consequently the training process results in a set of weights
{
|=1,2,…,}. To classify a feature vector
representing the walk of an individual, we compute the inner
product of its expanded feature vector with each of the weight
vectors. This results in a score sequence sm m=1,2,..,M. The
class label of the feature vector is then determined by
argmax(s).
5. EXPERIMENTAL RESULTS
In the following results we validate the proposed feature
extraction schemes on Gulf costumes and compare the results
against those obtained on casual costumes which are similar to
what is reported in the literature. Common to all of the results
to follow, we report the gait classification rate obtained from
training and testing the system with feature vectors of different
lengths according to the DCT cutoff as explained in Section 3
above. Unless otherwise stated, the classification results are
obtained using a least-squares classifier without polynomial
expansion.
We start the experimental results section by comparing between
different approaches to spatial domain feature extraction. In
Figure 3, the female and male datasets are mixed, this also
includes mixing different directions of walking (i.e. from left to
right and vice versa). For each individual, 75% of the walking
samples are used for training and 25% for testing. Hence the
testing data is unseen by the training model. The figure shows
that the Radon transformation with horizontal projections of
AP images results in the highest classification results. Intuitively
this makes sense because the horizontal projection represents
the shape of the accumulated motion from both the front and
the rear of the body of an individual. At a DCT cutoff of 60
coefficients, the classification results are very close to 100%.
This should not come as a surprise as similar results have been
reported in [3]. On the other hand the figure shows that Radon
transformation with vertical projection of the AP images results
in a very poor classification result. This is so because such
projections can only describe the height of the individual and
the sinusoidal-like motion of the head during the walk. Clearly
such features are not enough for identifying an individual.
Interleaving the feature vectors of both aforementioned
projection though results in an acceptable result as shown in the
figure. Feature extraction using zonal coding resulted is a
moderate classification results. This can be justified by the fact
that the AP images contain plenty of high frequencies, hence
describing such images whilst discarding most of the high
frequency content though zonal coding does not result in
accurate and precise feature vectors. Lastly, it is worth
mentioning that the above discussion applies equally for both
Gulf and casual customs. However in the latter scenario, the
classification scores resulting from the horizontal projections
are a bit more accurate. This is not a surprise as the Gulf
costume conceals some details of the body motion.
(a)
0
0.2
0.4
0.6
0.8
1
10 20 30 40 50 60
Classification rate
Length of feature vector
Horizontal Projection Vertical projection
Interleaved projection Zonal coding
(b)
Fig. 3. Comparison between various spatial feature extraction
approaches. (a) Gulf costume (b) Casual costume.
In Figure 4 we interleave the two directions of walking in terms
of system training and testing. That is we train the system on
one direction of walking and test it on the other direction.
Clearly this experiment is carried out in a cross-validation
manner and the average classification result is reported. In this
experiment 50% of the feature vectors belong to each direction
of walking hence the training to testing ratio is set as such. The
spatial feature extraction approach is Radon transformation
with horizontal projections. Clearly, the classification rates
based on different direction of walking results is less accurate as
shown in the figure, nonetheless, at a DCT cutoff of 60 a
classification rate of around 90% is achieved. The figure also
shows that with training and testing based on second order
reduced model polynomial expansion higher classification
results are obtained. At low DCT cutoffs the enhancement is
quite evident. However at higher dimensionality and due to the
low number of training samples per individual (4 in this
experiment), the matrix of the expanded feature vector matrix
becomes ill-conditioned thus affecting the matrix inverse
operation in the computation of the model weights. Again the
same discussion applies for both the Gulf and the casual
costume.
(a)
(b)
Fig. 4.Classification rates using different training approaches
based on the direction of walking. (a) Gulf costume (b) Casual
costume.
6. CONCLUSION
This paper proposed a solution for gait recognition using non-
western costumes. In particular, the work was considered with
Gulf customs for both genders. The proposed solution was also
tested on casual costumes and was shown to work as well. As
such the same system can be deployed for identifying
individuals with mixed customs without the need for a
customized solution for a particular custom. The paper
proposed to accumulate the prediction errors of consecutive
video images using an adaptive forward/backward prediction
scheme. This was needed to counter effect the relative motion
of the background objects. Once the motion is accumulated
into one or two images, spatial feature extraction was applied. It
was shown that the Radon transformation with horizontal
image projections result in precise and concise feature vectors
that are linearly separable. The experimental results revealed
that the proposed solution results in accurate classification rates
and works equally well for both of the aforementioned
costumes.
REFERENCES
[1] S.V. Stevenage, M.S. Nixon, and K. Vince, “Visual analysis
of gait as a cue to identity,” Applied Cognitive Psychology, vol.
13, pp. 513-526, Dec. 1999.
[2] G. Zhao, R. Chen, G. Liu, and H. Li, “Amplitude spectrum-
based gait recognition,” Proc. Int’l Conf. Automatic Face and
Gesture Recognition, pp. 23-28, 2004.
[3] S. Sarkar, P. Jonathon Phillips, Z. Liu, I. Robledo, P.
Grother, K. W. Bowyer, “The human id gait challenge problem:
data sets, performance, and analysis,” IEEE Transactions on
Pattern Analysis and Machine Intelligence, 27(2), pp. 162 – 177,
Feb. 2005
[4] Z.i Liu and S. Sarkar, “Improved gait recognition by gait
dynamics normalization”, IEEE Transactions on pattern
analysis and machine intelligence, (28)6, June 2006.
0
0.2
0.4
0.6
0.8
1
10 20 30 40 50 60
Classification rate
Length of feature vector
Horizontal Projection Vertical projection
Interleaved projection Zonal coding
0.4
0.6
0.8
1
10 20 30 40 50 60
Classification rate
Length of feature vector
Training dataset based on mixed directions of walking
Training dataset based on a different direction of walking
With 2nd order reduced model expansion
0.4
0.6
0.8
1
10 20 30 40 50 60
Classification rate
Length of feature vector
Training dataset based on mixed directions of walking
Training dataset based on a different direction of walking
With 2nd order reduced model expansion
[5] P.J. Phillips, S. Sarkar, I. Robledo, P. Grother, and K.
Bowyer, “The gait identification challenge problem: data sets
and baseline algorithm,” Proc. Int’l Conf. Pattern Recognition,
pp. 385-388, 2002.
[6] L. Wang, T. Tan, H. Ning, and W. Hu, “Silhouette analysis-
based gait recognition for human identification,” IEEE Trans.
Pattern Analysis and Machine Intelligence, 25(12), pp. 1505-
1518, Dec. 2003.
[7] S.D. Mowbry and M.S. Nixon, “Automatic Gait Recognition
via Fourier Descriptors of Deformable Objects,” Proc. Conf.
Audio Visual Biometric Person Authentication, pp. 566-573,
2003.
[8] Z. Liu and S. Sarkar, “Simplest representation yet for gait
recognition: averaged silhouette,” Proc. Int’l Conf. Pattern
Recognition, vol. 4, pp. 211-214, 2004.
[9] A. Johnson and A. Bobick, “A multi-view method for gait
recognition using static body parameters,” Proc. Int’l Conf.
Audio and Video-Based Biometric Person Authentication, pp.
301-311, 2001.
[10] X. Yang, Y. Zhou, T. Zhang, G. Shu, J. Yang, “Gait
recognition based on dynamic region analysis,” Signal
Processing, 88(9), pp. 2350-2356, September 2008.
[11] J. Wu, J. Wang, L. Liu, “Feature extraction via KPCA for
classification of gait patterns,” Human Movement Science,” pp.
393-411, 26(3), June 2007.
[12] N. Boulgouris, Z. Chi, “Human gait recognition based on
matching of body components,” Pattern
Recognition,” 40(6), pp. 1763-1770, June 2007.
[13] J. Han and B. Bhanu, “Statistical feature fusion for gait-
based human recognition,” Proc. IEEE Conf. Computer Vision
and Pattern Recognition, vol. 2, pp. 842-847, June 2004.
[14] Y. Liu, R. Collins, and Y. Tsin, “Gait sequence analysis
using frieze patterns,” Proc. European Conf. Computer Vision,
pp. 657-671, May 2002.
[15] L. Wang, T. Tan, H. Ning, and W. Hu, “Silhouette analysis-
based gait recognition for human identification,” IEEE
Transactions on pattern analysis and machine intelligence,
25(12), December 2003.
[16] M.-H. Cheng, M.-F. Ho, C.-L. Huang, “Gait analysis for
human identification through manifold learning and HMM,”
Pattern Recognition, 41(8), pp. 2541-2553, August 2008.
[17] T. Shanableh and K. Assaleh, “Telescopic vector
composition and polar accumulated motion residuals for
feature extraction in Arabic Sign Language recognition,”
EURASIP Journal on Image and Video Processing, vol. 2007,
Article ID 87929, 10 pages, 2007. doi:10.1155/2007/87929.
[18] K. Assaleh, and H. Al-Nashash,A Novel Technique for
the Extraction of Fetal ECG Using Polynomial Networks,”
IEEE Transactions on Biomedical Engineering, 52(6), pp. 1148
– 1152, June 2005.
... In [373], the authors compute gait moment image for gait representation which encodes the probability of an image at each keyframe. The authors in [374] propose a gait recognition system for gulf clothing using an accumulated prediction image. The technique proposed in [375] computes the binomial distribution of each pixel to obtain a gait probability image. ...
Article
Full-text available
Visual surveillance has exponentially increased the growth of security devices and systems in the digital era. Gait-based person identification is an emerging biometric modality for automatic visual surveillance and monitoring as the walking patterns highly correlate to the subject’s identity. The scientific research on person identification using gait has grown dramatically over the past two decades due to its several benefits. It does not require active collaboration from users and can be performed without their cooperation. It is difficult to be impersonated and identification can be validated from low-resolution videos and with simple instrumentation. This paper presents a comprehensive overview of the exiting techniques, their key stages, and recent developments in vision-based person identification using gait. We reviewed the historical research on gait locomotion and explain that how it is used to recognize the identity. The article summarizes the different types of features that have been proposed to encode the biomechanics of gait and also groups them into different categories and subcategories based upon the similarity in their implementation. We also present the impact of different covariate factors that affect the performance of gait recognition systems and also discuss the recent works to cope with these challenges. Furthermore, a comparison of the recognition accuracies reported by the existing algorithms to assess their performance under verification and identification mode is also presented. A detailed summary of publicly available vision-based gait databases is also provided. Finally, it offers insight into the challenges and open problems for future perspectives in the field of gait recognition that can help to set the directions for future research in this field.
... • Boulgouris and Chen [25] 2007 Radon Energy Image (REI) LDA nearest-neighbor • Liu and Zheng [115] 2007 Motion History Image (MHI) LDA nearest-neighbor • Ma et al. [133] 2007 Gait Moment Image (GMI) -nearest-neighbor • Yang et al. [193] 2008 Enhanced GEI (EGEI) Discriminative common vectors nearest neighbor • Chen et al. [31] 2009 Frame Difference Energy Image (FDEI) -HMM • Bashir et al. [15] 2009 Gait Flow Fields (GFF) PCA+LDA nearest-neighbor • Shanableh et al. [164] 2009 Accumulated Prediction Image (API) -Polynomial networks • Bashir et al. [14] 2010 Gait Energy Image (GEnI) PCA+LDA nearest-neighbor • Wang et al. [179,180] 2010 Chrono Gait Image (CGI) PCA+LDA nearest-neighbor • Zhang et al. [202] 2010 Active Energy Image (AEI) 2DLPP nearest neighbor • Mu et al. [142] 2010 C1Gait Discriminative locality alignment nearest-neighbor • Lam et al. [98] 2011 Gait Flow Image (GFI) LDA nearest neighbor • Roy et al. [160] 2012 Pose Energy Image (PEI) PCA+LDA nearest-neighbor • Hofmann and Rigoll [62] 2012 Gradient Histogram Energy Image (GHEI) PCA+LDA nearest-neighbor • Huang et al. [72] 2012 Shifted Energy Image (SEI) + Gait Structural Profile (GSP) LDA nearest neighbor • Liu et al. [119] 2012 Multiple HOG PCA + LDA nearest-neighbor • Jeevan et al. [81] 2013 Gait Pal and Pal Entropy (GPPE) PCA SVM • Hu et al. [67] 2013 LBP Flow -HMM • Lee et al. [101] 2014 Gait Probability Image (GPI) -minimum Kullback-Leibler (KL) • Kusakunniran [89] 2014 Histogram of Optical Flow (HOF) + Histogram of Oriented Gradient (HOG) -nearest-neighbor • Kusakunniran [88] 2014 Histogram of Optical Flow (HOF) + Histogram of Oriented Gradient (HOG) -SVM • Lee et al. [102] 2014 Time-sliced Averaged Motion History Image (TAMHI) -majority voting • Chen and Liu [33] 2014 Gait Differential Image (AGDI) 2DPCA nearest-neighbor • Arora and Srivastava [9] 2015 Gait Gaussian Image (GGI) -nearest neighbor • Arora et al. [10] 2015 Gradient Histogram Gaussian Image (GHGI) -nearest-neighbor • Lee et al. [103] 2015 Transient Binary Patterns (TBP) -majority voting • Luo et al. [130] 2015 Accumulated Frame Difference Energy Image (AFDEI) -nearest-neighbor • Arora et al. [8] 2015 Gait Information Image with Energy Feature/Sigmoid Feature (GII-EF/SF) -nearest-neighbor • Choudhury and Tjahjadi [40] 2016 Averaged Gait Key-phase Image (AGKI) PCA rotation forest ensemble • Chhatrala and Jandhav [35] 2016 Gabor Cosine Features (GCF) MLDA LGSR • Medikonda et al. [141] 2016 Generalized New Entropy (GNE) -SVM • Al et al. [3] 2017 Accumulated Flow Image (AFI) + Edge-Masked Active Energy Image (EMAEI) MPCA+LDA nearest-neighbor • Atta et al. [11] 2017 5/3 Gait Image (5/3GI) PCA nearest-neighbor • Chaurasia et al. [30] 2017 P RW D F GEI PCA+ Generalized LDA nearest-neighbor • Verlekar et al. [177] 2017 . It should be also noted that gait was combined with face in order to achieve higher accuracy [4,63,163]. ...
Article
Full-text available
Gait recognition has emerged as an attractive biometric technology for the identification of people by analysing the way they walk. However, one of the main challenges of the technology is to address the effects of inherent various intra-class variations caused by covariate factors such as clothing, carrying conditions, and view angle that adversely affect the recognition performance. The main aim of this survey is to provide a comprehensive overview of existing robust gait recognition methods. This is intended to provide researchers with state of the art approaches in order to help advance the research topic through an understanding of basic taxonomies, comparisons, and summaries of the state-of-the-art performances on several widely used gait recognition datasets.
Article
Full-text available
Gait Recognition is one of the latest and attractive biometric techniques, due to its potential in identification of individuals at a distance, unobtrusively and even using low resolution images. In this paper we focus on single lateral view gait recognition with various carrying and clothing conditions. Such a system is needed in access control applications whereby a single view is imposed by the system setup. The gait data is firstly processed using three gait representation methods as the features sources; Accumulated Prediction Image (API) and two new gait representations namely; Accumulated Flow Image (AFI) and Edge-Masked Active Energy Image (EMAEI). Secondly, each of these methods is tested using three matching classification schemes; image projection with Linear Discriminant Functions (LDF), Multilinear Principal Component Analysis (MPCA) with K-Nearest Neighbor (KNN) classifier and the third method: MPCA plus Linear Discriminant Analysis (MPCA + LDA) with KNN classifier. Gait samples are fed into the MPCA and MPCALDA algorithms using a novel tensor-based form of the gait images. This arrangement results into nine recognition sub-systems. Decisions from the nine classifiers are fused using decision-level (majority voting) scheme. A comparison between unweighted and weighted voting schemes is also presented. The methods are evaluated on CASIA B Dataset using four different experimental setups, and on OU-ISIR Dataset B using two different setups. The experimental results show that the classification accuracy of the proposed methods is encouraging and outperforms several state-of-the-art gait recognition approaches reported in the literature.
Conference Paper
Full-text available
We describe a new method for Automatic Gait Recognition based around the use of Fourier descriptors that model the periodic deformation of human gait. Fourier descriptors have been used successfully in the past to model the boundary of static or moving, rigid-bodied objects, but many objects actually deform in some way as they move. Here we use Fourier descriptors to model not only the object’s boundary, but also the spatio-temporal deformations under which the object’s boundary is subjected. We applied this new method to the Large Gait Database, compiled at the University of Southampton, and found that the Fourier descriptors obtained for each person appear to be unique and can be used for recognition. Successful recognition rates of over 85% were obtained from the Large Gait Database using only a small set of descriptors.
Article
Full-text available
Few researchers have investigated whether the timing of postevent information aspects the accuracy of children's reports of events they have experienced. In this study, four-year-olds dressed up in costumes and had their photographs taken. An unfamiliar adult spoke to the children about the event either a day (immediate condition) or a month (delayed condition) later, providing both accurate and misleading information about the staged event. When questioned five weeks after the event, children in a control group who had not received the review were more inaccurate answering focused questions than children who had been reminded of the event. A review a while after the event but shortly before the interview increased the amount of details recalled and this was not at the expense of accuracy. Mis-information was seldom reported spontaneously, although children in all groups acquiesced to leading questions in line with the misleading suggestions.
Conference Paper
Full-text available
A multi-view gait recognition method using recovered static body parameters of subjects is presented; we refer to these parameters as activity-specific biometrics. Our data consists of 18 subjects walking at both an angled and frontal-parallel view with respect to the camera. When only considering data from a single view, subjects are easily discriminated; however, discrimination decreases when data across views are considered. To compare between views, we use ground truth motion-capture data of a reference subject to find scale factors that can transform data from different viewsi nto a common frame ("walking-space"). Instead of reporting percent correct from a limited database, we report our results using an expected confusion metric that allows us to predict how our static body parameters filter identity in a large population: lower confusion yields higher expected discrimination power. We show that using motion-capture data to adjust vision data of different views to a common reference frame, we can get achieve expected confusions rates on the order of 6%.
Article
Full-text available
For humans, the ability to discriminate between, and to identify, others is paramount. The most obvious way this is accomplished is by means of face recognition. However, this is not the only method available. The present article reports on two experiments designed to see whether gait can be used as a reliable cue to identity. Experiment One showed that the human visual system was sophisticated enough to learn to identify six individuals on the basis of their gait signature under conditions of simulated daylight, simulated dusk and point-light displays. It thus appeared that gait-related judgements could be made, and furthermore, that these judgements were possible without reliance on shape information. Experiment Two suggested that even under adverse viewing conditions involving a single brief exposure, humans could identify a target from a ‘walking identity parade’ at greater than chance levels. These results emerged regardless of the lighting conditions, and were largely independent of the gender of the target walker. As such, the present results suggest that gait could be used as a reliable means of discriminating between individuals, and the importance of such an identity cue, in conditions in which the face is obscured, are discussed. Copyright © 1999 John Wiley & Sons, Ltd.
Article
Full-text available
This work introduces two novel approaches for feature extraction applied to video-based Arabic sign language recognition, namely, motion representation through motion estimation and motion representation through motion residuals. In the former, motion estimation is used to compute the motion vectors of a video-based deaf sign or gesture. In the preprocessing stage for feature extraction, the horizontal and vertical components of such vectors are rearranged into intensity images and transformed into the frequency domain. In the second approach, motion is represented through motion residuals. The residuals are then thresholded and transformed into the frequency domain. Since in both approaches the temporal dimension of the video-based gesture needs to be preserved, hidden Markov models are used for classification tasks. Additionally, this paper proposes to project the motion information in the time domain through either telescopic motion vector composition or polar accumulated differences of motion residuals. The feature vectors are then extracted from the projected motion information. After that, model parameters can be evaluated by using simple classifiers such as Fisher's linear discriminant. The paper reports on the classification accuracy of the proposed solutions. Comparisons with existing work reveal that up to 39% of the misclassifications have been corrected.
Article
Gait Energy Image (GEI) has been proved to be an effective identity signature in gait recognition. But previous approaches only treat this 2D image representation as a holistic feature and neglect the intrinsic dynamic characteristics of gait patterns. In this paper, we use variation analysis to obtain the dynamic region in GEI which reflects the walking manner of an individual. Based on this analysis, a dynamics weight mask is constructed to enhance the dynamic region and suppress the noises on the unimportant regions. The obtained gait representation called enhanced GEI (EGEI) is then represented in low dimensional subspace by Gabor-based discriminative common vectors analysis. We test the proposed approach on the USF HumanID Gait Database. Experimental results prove its effectiveness in terms of recognition rate.
Article
This paper presents a novel approach for gait recognition based on the matching of body components. The human body components are studied separately and are shown to have unequal discrimination power. Several approaches are presented for the combination of the results obtained from different body components into a common distance metric for the evaluation of similarity between gait sequences. A method is also proposed for the determination of the weighting of the various body components based on their contribution to recognition performance. Using the best performing of the proposed methods, improved recognition performance is achieved.
Conference Paper
With the increasing demands of visual surveillance systems, human identification at a distance has gained more attention from the researchers recently. Gait analysis can be used as an unobtrusive biometric measure to identify people at a distance without any attention of the human subjects. We propose a novel effective method for both automatic viewpoint and person identification by using only the silhouette sequence of the gait. The gait silhouettes are nonlinearly transformed into low-dimensional embedding by Gaussian process latent variable model (GPLVM), and the temporal dynamics of the gait sequences are modeled by hidden Markov models (HMMs). The experimental results show that our method has higher recognition rate than the other methods. (c) 2007 Elsevier Ltd. All rights reserved.
Article
Gait recognition is a process of identifying individuals by the way they walk. Gait is often used as a unobstrusive biometric offering the possibility to identify people at a distance without any interaction or co-operation with the subject. This paper presents a novel method for both automatic viewpoint and person identification using only the silhouette sequence of gait. The gait silhouettes are nonlinearly transformed into low dimensional embedding and the dynamics in time-series images are modeled by HMM in the corresponding embedding space. The experimental results demonstrate that the proposed algorithm is an encouraging progress for the gait analysis research.