ArticlePDF Available

Geometric Approach for Human Emotion Recognition using Facial Expression

Authors:

Abstract

Paper contains emotion recognition system based on facial expression using Geometric approach. A human emotion recognition system consists of three steps: face detection, facial feature extraction and facial expression classification. In this paper, we used an anthropometric model to detect facial feature points. The detected feature points are group into two class static points and dynamic points. The distance between static points and dynamic points is used as a feature vector. Distance changes as we track these points in image sequence from neutral state to corresponding emotion. These distance vectors are used for input to classifier. SVM (Support Vector Machine) and RBFNN (Radial Basis Function Neural Network) used as classifier. Experimental results shows that the proposed approach is an effective method to recognize human emotions through facial expression with an emotion average recognition rate 91 % for experiment purpose the Cohn Kanade databases is used.
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
17
Geometric Approach for Human Emotion
Recognition using Facial Expression
S. S. Bavkar
Assistant Professor
VPCOE Baramati
J. S. Rangole
Assistant Professor
VPCOE Baramati
V. U. Deshmukh
Assistant Professor
VPCOE Baramati
ABSTRACT
Paper contains emotion recognition system based on facial
expression using Geometric approach. A human emotion
recognition system consists of three steps: face detection,
facial feature extraction and facial expression classification. In
this paper, we used an anthropometric model to detect facial
feature points. The detected feature points are group into two
class static points and dynamic points. The distance between
static points and dynamic points is used as a feature vector.
Distance changes as we track these points in image sequence
from neutral state to corresponding emotion. These distance
vectors are used for input to classifier. SVM (Support Vector
Machine) and RBFNN (Radial Basis Function Neural
Network) used as classifier. Experimental results shows that
the proposed approach is an effective method to recognize
human emotions through facial expression with an emotion
average recognition rate 91 % for experiment purpose the
Cohn Kanade databases is used.
Keywords
Geometric Method, Anthropometric model, SVM, RBFNN
and LK Tracker
1. INTRODUCTION
Recently there has been a growing interest in improving the
interaction between humans and computers. It is argued that
to achieve effective human-computer intelligent interaction,
there is a need for the computer to interact naturally with the
user, similar to the way humans interact. Humans interact
with each other mostly through speech, but also through body
gestures to emphasize a certain part of speech and/or display
of emotions. Emotions are displayed by visual, vocal and
other physiological means. There is more and more evidence
appearing that shows that emotional skills are part of what is
called ‘intelligence’. One of the most important ways for
humans to display emotions is through facial expressions.
Mehrabian [1] points out that 7% of human communication
information are communicated by linguistic language (verbal
part), 38% by paralanguage (vocal part) and 55% by facial
expression. Therefore, facial expressions are the most
important information for emotional perception in face to face
communication. The emotion recognition system can be
broadly classified in two methods: Appearance based
(Texture) and Geometric based. Texture-based method model
local texture around a given feature point [2][3], for example
the pixel values in a small region around a mouth corner.
Geometric-based methods regard all facial feature points as a
shape [4], which is learned from a set of labeled faces and try
to find the proper shape for any unknown face.
The remainder of this paper is organized as follows: section
II, specify the related work, and then in section III facial
feature point localization system is presented. Our facial
expression recognition system is presented in section IV,
where we specify the feature extraction method and the
recognition approach. In section V, we present the obtained
experimental results. Finally, conclusion is presented in
section VI.
2. RELATED WORK
A Neural Network (NN) is employed to perform facial
expression recognition in [5]. The features used can be either
the geometric positions of a set of fiducial points on a face or
a set of multi-scale and multi-orientation Gabor wavelet
coefficients extracted from the facial image at the fiducial
points. The recognition is performed by a two layer
perceptron NN. A Convolutional NN was used in [6]. The
system developed is robust to face location changes and scale
variations. Feature extraction and facial expression
classification were performed using neuron groups, having as
input a feature map and properly adjusting the weights of the
neurons for correct classification. A method that performs
facial expression recognition is presented in [7].
There are several works in the field of facial expression
recognition using feature points. The added value of our work
is the modelization of muscle contraction using the variation
of muscle distances relative to the neutral state. The previous
studies on facial feature modelization are based on:
1) Points displacements [8]
2) Facial points coordinate [9],
3) Distances between points but it is based on the deformation
of the facial contour [10],
4) Deformation of the shape from the neutral state but not on
the contraction of facial muscles [11].
3. FACIAL FEATURE POINTS
LOCALIZATION
For facial feature point localization, a anthropometric model
is developed on detected face. Corners of the eyes, corners of
the eyebrows, corners and outer mid points of the lips, corners
of the nostrils, tip of the nose, and the edge of the face are
localized as facial feature points. After detecting the facial
feature points the first step is to track these points throughout
the image sequence. Currently, facial feature point’s
localization is usually carried out by manually labeling the
points.
3.1 Face Detection
Face detection is the first step in our facial expression
recognition system to localize the face in the image. A real-
time face detector proposed in [12] which represents an
adapted version of the original Viola-Jones face detector (Fig.
1-a). The next step in the automatic facial point localization is
to determine the coarse position for each point. To achieve
this, we develop a fully automatic method using an
anthropometric model.
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
18
Anthropometry is a biological science that deals with the
measurement of the human body and its different parts. Data
obtained from anthropometric measurement informs a range
of enterprises that depend on knowledge of the distribution of
measurements across human populations. After carefully
performing anthropometric measurement on Cohn Kanade
database [13], we have been able to build an anthropometric
model of human face that can be used in localizing facial
feature points from face images. The landmarks points that
have been used in our face anthropometric model for facial
feature localization are represented in Fig. 1-e. It has been
observed from the statistics of proportion evolved during our
initial observation that location of these points (P1 to P38) can
be obtained from: The distance D measured between eyes axis
EA and mouth axis MA (Fig. 4-c); Face symmetry axis SA as
reference for x position of points.
The facial feature point’s localization has been found from
proportional constants in (Table I) using distance between
eyes axis, mouth axis, symmetry axis and the face box center
YC as the principle parameter of measurement. To localize
feature points on face we have to find out MA, SA, EA and
center of face box.
(a) Face Detection
(b) Eye Axis localization
(c) Mouth Axis Localization
(d) Points Localization using Anthropometric model
Fig. 1: Facial feature points detection outline
3.2 Main Axis Localization
The key step of our anthropometric model realization is facial
features axis localization. Fig. 4-c shows the three main axis
which passed by facial features: two horizontal axis for mouth
(MA) and eyes (EA) and vertical axis passed by nose which
give the symmetry of the face (SA). YC is the face box center.
3.1.1 Eyes axis localization
It is determined by the maximum of the projection curve
which has a high gradient. First, we calculate the gradient of
the image I (corresponds to the face rectangle extracted by the
Viola-Jones detector):
(1)
Ix corresponds to the differences in the x direction. The
spacing between points in each direction is assumed to be one.
Computing the absolute gradient value in each line is given
by:
n
y
yx
x
Ix
x
HI
1
),()(
(2)
Then, we find the maximum value which corresponds to the
line contains eyes (Fig. 1-b). This line corresponds to many
transitions: skin to sclera, sclera to iris, iris to pupil and the
same thing for the other side (high gradient).
Table 1. Proportion of facial feature points positions
measured from subjects of different geographical
territories
Point
X position
Y position
P1
P2
P3
P4
P5
P6
P7
P8
P9
P10
P11
P12
P13
P14
P15
P16
P17
P18
P19
P20
P21
P22
P23
P24
P25
P26
P27
P28
P29
P30
P31
P32
P33
P34
P35
P36
P37
P38
SA-0.91*D
SA-0.58*D
SA-0.26*D
SA+0.26*D
SA+0.58*D
SA+0.91*D
SA+1.04*D
SA+0.71*D
SA+0.52*D
SA+0.13*D
SA-0.13*D
SA-0.52*D
SA-0.71*D
SA-1.04*D
SA-0.06*D
SA+0.06*D
SA-0.78*D
SA-0.58*D
SA-0.26*D
SA+0.26*D
SA+0.58*D
SA+0.78*D
SA-0.26*D
SA+0.26*D
SA-0.65*D
SA-0.32*D
SA
SA+0.32*D
SA+0.45*D
SA+0.32*D
SA+0.19*D
SA
SA-0.19*D
SA-0.32*D
SA-0.74*D
SA-0.58*D
SA+0.58*D
SA+0.74*D
YC-0.91*D
YC-1.17*D
YC-1.3*D
YC-1.3*D
YC-1.17*D
YC-0.91*D
YC
YC+0.91*D
YC+1.17*D
YC+1.3*D
YC+1.3*D
YC+1.17*D
YC+0.91*D
YC
YC-0.26*D
YC-0.26*D
YC-0.52*D
YC-0.58*D
YC-0.52*D
YC-0.52*D
YC-0.58*D
YC-0.52*D
YC+0.32*D
YC+0.32*D
MA
MA-0.1*D
MA-0.15*D
MA-0.1*D
MA
MA+0.1*D
MA+0.13*D
MA+0.15*D
MA+0.13*D
MA+0.1*D
YC-0.26*D
YC-0.39*D
YC-0.39*D
YC-0.26*D
I
x
I
Ix
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
19
3.2.2 Mouth axis localization
To locate the mouth axis, we first define a Region of Interest
(ROI) of the mouth to be the horizontal strip whose top is at
0.67*R from the face bounding box top and has a width equal
to 0.25*R. This strip is located around the median of the
bounding box of the face with a width of 0.1*R, where R is
the side face box. Intensity information is used to locate
mouth axis, minimum intensity at mouth.
3.2.3 Symmetry axis localization
It is a vertical line which divides the frontal face in two equal
sides. To locate the symmetry axis, we first define a ROI of
the nose. Since knowing the location of eyes axis (EA) and
mouth axis (MA), we define the nose region to be the vertical
strip whose top is the eyes axis and has a height equal to D.
This strip is located around the median of the bounding box of
the face with a width of 10% of the face windows width.
Analysis of the gray level vertical projection of the ROI of the
nose shows that the maximum of the projection curves
corresponds to symmetry axis
Experimental results show that the extraction of the facial
feature points using anthropometric model gives a good
location independently to skin color and illumination changes
(Fig. 2).
4. FEATURE POINTS TRACKING
To track the localized feature Lucas Kanade Optical flow
tracker is used [14]. Optical flow is defined as an apparent
motion of image brightness. Two main assumptions can be
made Su & Hsieh (2007):
1. Brightness I(x, y, t) smoothly depends on coordinates x, y
in greater part of the image.
2. Brightness of every point of a moving or static object
does not change in time [14].
Fig. 2: Facial feature point’s localization
Let some object in the image the image or some point of an
object, move and after time dt the object displacement is (dx ,
dy). Using Taylor series for brightness I(x, y, t)
(3)
Then according to assumption 2:
(4)
And
(5)
Usually above equation called optical flow constraint
equation, where
Are component of optical flow field in x and y coordinates
respectively.
Calculate optical flow returns to calculate for each point in the
image the following equation:
(6)
However, the above equation cannot determine with a single
way the optical flow. The indetermination of optical flow due
to the absence of global constraint in precedent equations,
only gradients which are local measures are taken into
account. Lucas and Kanade have added new constraints to
ensure the uniqueness of the solution. The method of Lucas
and Kanade consists to find point location in next image by
applying a calculation of least squares to minimize constraint.
They define a pre-neighborliness, and they optimize the above
equation to give solution of the following system for n points:
(7)
5. FACIAL EXPRESSION
RECOGNITION
Our approach uses a feature based representation of facial data
for SVM and RBFNN classifier. It classifies single images
taken from an image sequence with respect to six basic
emotions of Ekman [15] happy, fear, disgust, anger, sadness,
surprise and neutral state. Our work is based on the facial
features deformation compared to the neutral state.
Fig. 3: Facial features point’s detection and tracking in
image sequence.
...),,(),,(
dt
t
I
dy
y
I
dx
x
I
tyxIdttdyydxxI
0...
dt
t
I
dy
y
I
dx
x
I
u
dt
dx
v
dt
dy
y
I
v
x
I
u
t
I
..
)(
.
.
.
)(
.
.
.
)(
.
)()(
..
..
..
)()(
..
..
..
)()( 1
11
n
i
nn
ii
p
x
I
p
x
I
p
x
I
v
u
p
y
I
p
x
I
p
y
I
p
x
I
p
y
I
p
x
I
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
20
5.1 Coding
In fact, the human facial expressions originate from the
movements of facial muscles beneath the skin. Thus, we
represent each facial muscle by a pair of key points, namely
dynamic point and fixed point. As shown in Fig. 4-a, the
dynamic points can be moved during an expression, while
Fig. 4-b shows the fixed points which cannot be moved during
a facial expression (face edge, nose root and outer corners of
the eyes).
Further, each facial muscle is represented by a distance (Fig.
4-d), as: Eyebrows motions are described by the distances
from D1 to D7, Eyes motions are described by the distances D8
and D9, Nose motions are described by the distances D10 and
D11, Mouth motions are described by the distances from D12
to D21. These distances are calculated by Euclidean distance
formula.
(8)
The used method to encode a facial expression takes into
consideration all distances Di variations of each muscle during
the sequence. If DT = (d1, d2... di... d21) is the vector
parameters extracted from a video sequence at the moment T.
(a) (b) (c) (d)
Fig. 4: (a) Dynamic Points, (b) Static Points, (c)
Principal Axis, (d) Facial Distances
D is the distance variation from the first image (a neutral
expression), where:
(9)
Where D0i is the ith distance of the neutral state, with i ϵ [1,
21]. After extracting distance feature vector from static and
dynamic points, they are normalized in between 0 to 1.
5.2 Support Vector Machine
After the extraction of the necessary information from the
facial expression, we have trained a statistical classifier
Support vector machine SVM [16]. Support Vector Machine
makes binary decisions.
There are a number of methods for making multiclass
decisions with a set of binary classifiers. The simplest strategy
is to train 1 versus all remaining, but this method gives poor
results. To overcome this we adapted 1 versus 1 strategy. We
trained test emotion with all possible combination of other
pair. We get such 15 pair for six basic emotions. Then the
polling method is used the class having maximum voting, the
test emotion belongs to that class. In general, the RBF kernel
is a reasonable first choice. This kernel nonlinearly maps
samples into a higher dimensional space so it, unlike the
linear kernel, can handle the case when the relation between
class labels and attributes is nonlinear. In addition, the
sigmoid kernel behaves like RBF for certain parameters. The
second reason is the number of hyper parameters which
increase the complexity of model selection. The polynomial
kernel has more hyper parameters than the RBF kernel. The
rbf kernel is as follow.
Di
Fig. 5: Facial expression recognition system
5.3 Radial Basis Function Neural Network
The basic architecture for a RBF is a 3-layer network; input
layer is simply a fan-out layer and does no processing. The
second or hidden layer performs a non-linear mapping from
the input space into a (usually) higher dimensional space in
which the patterns become linearly separable. The final layer
therefore performs a simple weighted sum with a linear
output. If the RBF network is used for function approximation
(matching a real number) then this output is fine. However, if
pattern classification is required, then a hard-limiter or
sigmoid function could be placed on the output neurons to
give 0 or 1 output values. The unique feature of the RBF
network is the process performed in the hidden layer [17]. The
idea is that the patterns in the input space form clusters. If the
centres of these clusters are known, then the distance from the
cluster centre can be measured. Furthermore, this distance
measure is made non-linear, so that if a pattern is in an area
that is close to a cluster centre it gives a value close to 1.
Beyond this area, the value drops dramatically. The notion is
that this area is radically symmetrical around the cluster
centre, so that the non-linear function becomes known as the
radial-basis function. The most commonly used radial-basis
function is:
Where σ is the spread parameter of the Gaussian functions, r
is the distance from the cluster centre. The distance measured
from the cluster centre is usually the Euclidean distance. For
each neuron in the hidden layer, the weights represent the co-
ordinates of the centre of the cluster. Therefore, when that
neuron receives an input pattern, x, the distance is found using
the following equation.
Number of neurons is 250 and RBF spread value is 250 used
in this system.
021
21
001
1,....,.... D
D
D
D
D
D
D
i
i
2
21
2
21 )()( yyxxD
Decision
Function
Model
Facial
Expression
Test
Expression
1bxw if1
1bxw if1
)(
xf
i
ixx
xxK exp),(
)
2
exp()( 2
2
r
r
n
iijij wxr 1
2
)(
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
21
6. RESULT AND DISCUSSION
For the evaluation of work, we have used Cohn-Kanade [13]
databases. The systems will implement on MATLAB 7.5
version. Fig. 5 describes our facial expression recognition
system. Performance of system is checked using SVM and
RBFNN classifier.
The experiment performed using cross validation in order to
compute the accuracy. In Cohn Kanade database contain 50
emotions for each class, so initially we used 40 emotions from
each class for training and remaining 10 emotions for testing.
Then cross validation is used to increase the database size.
Totally 50 emotions from each class are used as testing
purpose. Table 2 and Table 3 represents confusion matrix of
different emotions using SVM and RBFNN classifier
respectively on Cohn-Kanade database. This matrix shows the
effectiveness of classification methods with the SVM. For our
classifier, we achieve the recognition rate averagely 91%. Fig.
6 shows the recognition rate for different expression using
SVM classifier.
Table 2. Emotion confusion matrix using SVM classifier
on Cohn kanade database
Anger
Disgust
Fear
Happy
Sad
Surprise
Anger
47
0
2
0
4
0
Disgust
0
41
1
5
2
1
Fear
1
1
47
0
0
0
Happy
0
6
0
45
0
0
Sad
2
2
0
0
44
0
Surprise
0
0
0
0
0
49
Average Recognition rate=91%
Table 3. Emotion confusion matrix using RBFNN
classifier on Cohn kanade database
Anger
Disgust
Fear
Happy
Sad
Surprise
Anger
47
0
2
0
5
0
Disgust
1
42
1
5
2
1
Fear
1
0
45
1
0
0
Happy
0
5
0
44
0
0
Sad
1
3
2
0
43
0
Surprise
0
0
0
0
0
49
Average Recognition rate=90%
Fig. 6: Recognition rate for different expression using
SVM Classifier
7. CONCLUSION AND FUTURE SCOPE
In this paper an automatic approach to emotion recognition
based on facial expression analysis presented. To detect the
face in the image, we have used the face detector of Viola
Jones which is fast and robust to illumination condition. For
feature point’s extraction, we have developed anthropometric
models which suppose that the positions of facial feature
points are proportional to the vertical distance between eyes
and mouth. Then that feature points are tracked throughout the
image sequence. The proposed method gives good results,
when tested on images from the Cohn-Kanade database, under
various illuminations. The tracking of the detected points have
been realized with Lucas-Kanade algorithm. Variations
distance vector was used as a descriptor of the facial
expression. This vector is the input of SVM and RBFNN
classifier. Emotion recognition rates about 91 % were
achieved in real time.
Proposed combination method for feature extraction does not
extract features parameters properly if there are hairs on face
area. Therefore in future an attempt can be made to develop
hybrid approach for facial feature extraction and recognition
accuracy can be further improved using NN approach and
hybrid approach such as ANFIS. An attempt can also be made
for recognition of other database images or images captured
from camera
8. ACKNOWLEDGMENTS
I would like to acknowledge Principal of Vidya Pratishthan’s
college of Engineering for providing research platform and
Head and faculty members of the Electronics department for
providing necessary support and valuable guidance for
presenting this research paper.
9. REFERENCES
[1] A. Mehrabian, “Communication without Words,”
Psychology Today, Vo.1.2, No.4, pp 53-56, 1968.
[2] Z. Xin, X. Yanjun, and D. Limin. “Locating facial
features with color information”. IEEE International
Conference on Signal Processing, 2 :889892, 1998.
[3] Shishir Bashyal, Ganesh K. Venayagamoorthy,
“Recognition of facial expressions using Gabor wavelets
and learning vector quantization”. Elsevier Engineering
Applications of Artificial Intelligence 21 (2008) 1056
1064.
[4] F. ABDAT, C. MAAOUI and A. PRUSKI, “Human-
computer interaction using emotion recognition from
facial expression” 2011 UKSim 5th European
Symposium on Computer Modeling and Simulation,
978-0-7695-4619, 2011.
[5] Z. Zhang, M. Lyons, M. Schuster, and S. Akamatsu,
“Comparison between geometry-based and Gabor-
wavelets-based facial expression recognition using multi-
layer perceptron”. in Proceedings of the 3rd IEEE
International Conference on Automatic Face and Gesture
Recognition, Nara Japan, 14-16 April 1998, pp. 454.459.
[6] B. Fasel, “Multiscale facial expression recognition using
Convolutional neural networks”. IDIAP, Tech. Rep.,
2002.
[7] M. Matsugu, K. Mori, Y.Mitari, and Y. Kaneda, “Subject
independent facial expression recognition with robust
face detection using a Convolutional neural network”.
Neural Networks, vol. 16, no. 5-6, pp. 555.559, June-July
2003.
47
0
2
0
4
0
0
41
1
5
2
1
1
1
47
0
0
0
0
6
0
45
0
0
2
2
0
0
44
0
0
0
0
0
0
49
Angry
Disgust
Fear
Happy
Sad
Surprise
Angry
Disgust
Fear
Happy
Sad
Surprise
International Journal of Computer Applications (0975 8887)
Volume 118 No.14, May 2015
22
[8] I. Cohen Y. Sun T. Gevers N. Sebe, M.S. Lew and T.S.
Huang. “Authentic facial expression analysis”. Proc.
IEEE Int’l Conf. Automatic Face and Gesture
Recognition (AFGR), 2004.
[9] J. Bailonson, E. Pontikakisb, I. Maussc, J. Grossd, M. Ja-
bone, C. Hutchersond, C. Nassa, and O. Johnf. Real-time
classification of evoked emotions using facial feature
tracking and physiological responses. International
Human Computer Studies, 66 :303317, 2008.
[10] Koen van de Sande Roberto Valenti Aitor Azcarate,
Felix Hageloh. Automatic facial emotion recognition,
2005.
[11] P.W. Yuille, A.L.and Hallinan and D.S. Cohen. “Feature
extraction from faces using deformable templates”.
International Journal of Computer Vision, 8:99111,
[12] P. Viola and M. Jones. “Robust real-time object
detection”. 2nd international workshop on statistical and
computational theories of vision - modeling, learning,
computing, and sampling”. canada, 2001.
[13] T. Kanade, J. F. Cohn, and Y. Tian. “Comprehensive
database for facial expression analysis”. Fourth IEEE
International Conference on Automatic Face and Gesture
Recognition Grenoble France, FG’00:46–53, 2000.
[14] J.Y. Bouguet. “Pyramidal implementation of the Lucas
kanade feature tracker”. Intel Corporation,
Microprocessor Research Labs, 2000.
[15] P. Ekman. Emotion in the human face. Cambridge
University Press, 1982.
[16] S. Gunn, “Support vector machines for classification and
regression”. Image Speech and Intelligent System Group,
Univ. of Southampton MP-TR-98-05, 1998.
[17] D. T. Lin. (2006). Human facial expression recognition
using hybrid network of PCA and RBFN, Lecture Notes
on Computer Science 4132, 624633.
IJCATM : www.ijcaonline.org
... In geometry-based feature extraction, the location and shape of facial areas are considered which necessitates precise facial features extraction. In [19] [20]. They have used a mathematical model on various locations on the face in an automated manner. ...
Chapter
Full-text available
Individuals suffering from Autism Spectrum Disorder find it challenging to perceive basic human emotions, which deters their communication capabilities. Given this difficulty, we proposed a Kawaii-engineered framework for Individuals with Autism by developing a Machine Learning pipeline using multilabel classification algorithms to identify the emotions from a video. We experimented on two datasets OMGE and DIAEMO. Both datasets have videos of the duration of approximately one minute. After pre-processing, facial expressions and audio content-based features were extracted. Multilabel Classification algorithm like Instance Based Learning by Logistic Regression for Multi-Label Learning, Multi Label k- Nearest Neighbour, Binary Relevance k- Nearest Neighbour, Random k- Label Sets, and Calibrated Label Ranking were used and for sampling, 5-fold Cross-Validation and Leave One Out Cross Validation were employed for both the datasets. We observed that LOOCV based sampling strategy gave the best results for both the datasets with CLR classifier using Gaussian kernel.
... Geometric-based methods are to obtain the feature vector representing the facial expression by taking into account the positions and shapes of the turning points (eye, mouth, eyebrow, nose) on the face [14]. There are many geometric-based studies in the FER field [15][16][17][18][19][20][21]. ...
Article
Full-text available
Facial expressions, which are important social communication tools in our daily life, provide important information about the mental state of people. Research is being done to obtain this information accurately. The importance of these researchs in the field of human-computer interaction is increasing. Many methods have been used for the recognition of universal facial expressions such as neutral, happiness, surprise, sadness, anger, disgust, and fear by intelligent systems with high accuracy. Emotion recognition is an example of difficult classification due to factors such as ambient light, age, race, gender, and facial position. In this article, a 3-stage system is proposed for emotion detection from facial images. In the first stage, the CNN-based network is trained with the Fer+ dataset. The Binary Particle Swarm Optimization algorithm is applied for feature selection to the feature vector in the fully connected layer of the CNN network trained in the second stage. Selected features are classified by Support Vector Machine. The performance of the proposed system has been tested with the Fer+ dataset. As a result of the test, 85.74% accuracy was measured. The results show that the combination of BPSO and SVM contributes to the classification accuracy and speed of the FER+ dataset.
Article
Full-text available
Facial expression recognition has many potential applications wh ich have attracted the attention of researchers in the last d ecade. Feature extract ion is one important step in exp ression analysis which contributes toward fast and accurate expression recogn ition. Happy, surprise, disgust, sad, anger and fear facial expressions Emot ion are of facial recognition. Facial expressions are mo st commonly used for interpretation of human emot ion. There is a range of different emot ions in two categories: positive emo t ion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emot ion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Full-text available
Facial expression recognition has many potential applications wh ich have attracted the attention of researchers in the last d ecade. Feature extract ion is one important step in exp ression analysis which contributes toward fast and accurate expression recogn ition. Happy, surprise, disgust, sad, anger and fear facial expressions Emot ion are of facial recognition. Facial expressions are mo st commonly used for interpretation of human emot ion. There is a range of different emot ions in two categories: positive emo t ion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emot ion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Full-text available
Facial expression recognition has many potential applications wh ich have attracted the attention of researchers in the last d ecade. Feature extract ion is one important step in exp ression analysis which contributes toward fast and accurate expression recogn ition. Happy, surprise, disgust, sad, anger and fear facial expressions Emot ion are of facial recognition. Facial expressions are mo st commonly used for interpretation of human emot ion. There is a range of different emot ions in two categories: positive emo t ion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emot ion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Full-text available
Facial expression recognition has many potential applications which have attracted the attention of researchers in the last decade. Feature extraction is one important step in expression analysis which contributes toward fast and accurate expression recognition. Happy, surprise, disgust, sad, anger and fear facial expressions Emotion are of facial recognition. Facial expressions are most commonly used for interpretation of human emotion. There is a range of different emotions in two categories: positive emotion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emotion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Facial expression recognition has many potential applications which have attracted the attention of researchers in the last decade. Feature extraction is one important step in expression analysis which contributes toward fast and accurate expression recognition. Happy, surprise, disgust, sad, anger and fear facial expressions Emotion are of facial recognition. Facial expressions are most commonly used for interpretation of human emotion. There is a range of different emotions in two categories: positive emotion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emotion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Full-text available
Facial expression recognition has many potential applications which have attracted the attention of researchers in the last decade. Feature extraction is one important step in expression analysis which contributes toward fast and accurate expression recognition. Happy, surprise, disgust, sad, anger and fear facial expressions Emotion are of facial recognition. Facial expressions are most commonly used for interpretation of human emotion. There is a range of different emotions in two categories: positive emotion and non-positive emotion. There are four types of generally using system: Face detection, extraction, Classification and recognition. In Existing system it is not so much identify exact emotion of a person. In this proposed taking the large scale image, hybrid extraction feature and ANN classification of frame based expression recognition try to detect facial expression detection and emotion detection for positive and non-positive images also design robust.
Article
Facial emotion recognition (FER) is the task of recognising human emotions from images and videos. Communicating through facial emotions is a kind of non-verbal communication and it reflects a person's inner thoughts and mental states. In the present study, various existing geometric and appearance based feature extraction techniques used in FER are reviewed in tabular form. The main motive of this paper is to analyse the performance of these techniques on the bases of accuracy on different datasets like JAFFE, CK+, CK and MMI. After extensive research on feature extraction techniques for FERS, it is found that the appearance feature-based techniques achieved maximum accuracy and more favourable as compared to geometric feature-based techniques. Finally, the paper concludes with the various challenges encountered for feature extraction in the field of FERS which need to be addressed in the future.
Article
Facial emotion recognition (FER) is the task of recognising human emotions from images and videos. Communicating through facial emotions is a kind of non-verbal communication and it reflects a person's inner thoughts and mental states. In the present study, various existing geometric and appearance based feature extraction techniques used in FER are reviewed in tabular form. The main motive of this paper is to analyse the performance of these techniques on the bases of accuracy on different datasets like JAFFE, CK+, CK and MMI. After extensive research on feature extraction techniques for FERS, it is found that the appearance feature-based techniques achieved maximum accuracy and more favourable as compared to geometric feature-based techniques. Finally, the paper concludes with the various challenges encountered for feature extraction in the field of FERS which need to be addressed in the future.
Article
Full-text available
Recognizing human facial expression and emotion by computer is an interesting and challenging prob- lem. In this paper we present a system for recogniz- ing emotions through facial expressions displayed in live video streams and video sequences. The system is based on the Piecewise Bezier Volume Deforma- tion tracker (18) and has been extended with a Haar face detector to initially locate the human face au- tomatically. Our experiments with Naive Bayes and the Tree-Augmented-Naive Bayes (TAN) classifiers in person-dependent and person-independent tests on the Cohn-Kanade database (1) show that good classi- fication results can be obtained for facial expression recognition.
Conference Paper
Full-text available
Within the past decade, significant effort has occurred in developing methods of facial expression analysis. Because most investigators have used relatively limited data sets, the generalizability of these various methods remains unknown. We describe the problem space for facial expression analysis, which includes level of description, transitions among expression, eliciting conditions, reliability and validity of training and test data, individual differences in subjects, head orientation and scene complexity, image characteristics, and relation to non -verbal behavior. We then present the CMU-Pittsburgh AU-Coded Face Expression Image Database, which currently includes 2105 digitized image sequences from 182 adult subjects of varying ethnicity, performing multiple tokens of most primary FACS action units. This database is the most comprehensive test-bed to date for comparative studies of facial expression analysis.
Article
Automatic face analysis has to cope with pose and lighting variations. Especially pose variations are difficult to tackle and many face analysis methods require the use of sophisticated normalization procedures. We propose a data-driven face analysis approach that is not only capable of extracting features relevant to a given face analysis task, but is also robust with regard to face location changes and scale variations. This is achieved by deploying convolutional neural networks. We show that the use of multi-scale feature extractors and whole-field feature map summing neurons allow to improve facial expression recognition results, especially with test sets that feature scale, respectively, translation changes.
Article
Facial expression recognition has potential applications in different aspects of day-to-day life not yet realized due to absence of effective expression recognition techniques. This paper discusses the application of Gabor filter based feature extraction in combination with learning vector quantization (LVQ) for recognition of seven different facial expressions from still pictures of the human face. The results presented here are better in several aspects from earlier work in facial expression recognition. Firstly, it is observed that LVQ based feature classification technique proposed in this study performs better in recognizing fear expressions than multilayer perceptron (MLP) based classification technique used in earlier work. Secondly, this study indicates that the Japanese Female Facial Expression (JAFFE) database contains expressers that expressed expressions incorrectly and these incorrect images adversely affect the development of a reliable facial expression recognition system. By excluding the two expressers from the data set, an improvement in recognition rate from 87.51% to 90.22% has been achieved. The present study, therefore, proves the feasibility of computer vision based facial expression recognition for practical applications like surveillance and human computer interaction.
Article
There is a growing trend toward emotional intelligence in human–computer interaction paradigms. In order to react appropriately to a human, the computer would need to have some perception of the emotional state of the human. We assert that the most informative channel for machine perception of emotions is through facial expressions in video. One current difficulty in evaluating automatic emotion detection is that there are currently no international databases which are based on authentic emotions. The current facial expression databases contain facial expressions which are not naturally linked to the emotional state of the test subject. Our contributions in this work are twofold: first, we create the first authentic facial expression database where the test subjects are showing the natural facial expressions based upon their emotional state. Second, we evaluate the several promising machine learning algorithms for emotion detection which include techniques such as Bayesian networks, SVMs, and decision trees.
Article
We present automated, real-time models built with machine learning algorithms which use videotapes of subjects’ faces in conjunction with physiological measurements to predict rated emotion (trained coders’ second-by-second assessments of sadness or amusement). Input consisted of videotapes of 41 subjects watching emotionally evocative films along with measures of their cardiovascular activity, somatic activity, and electrodermal responding. We built algorithms based on extracted points from the subjects’ faces as well as their physiological responses. Strengths of the current approach are (1) we are assessing real behavior of subjects watching emotional videos instead of actors making facial poses, (2) the training data allow us to predict both emotion type (amusement versus sadness) as well as the intensity level of each emotion, (3) we provide a direct comparison between person-specific, gender-specific, and general models. Results demonstrated good fits for the models overall, with better performance for emotion categories than for emotion intensity, for amusement ratings than sadness ratings, for a full model using both physiological measures and facial tracking than for either cue alone, and for person-specific models than for gender-specific or general models.
Conference Paper
This paper describes emotion recognition system based on facial expression. A fully automatic facial expression recognition system is based on three steps: face detection, facial characteristic extraction and facial expression classification. We have developed an anthropometric model to detect facial feature points combined to Shi&amp, Thomasi method. The variations of 21 distances which describe the facial features deformations from the neutral face, were used to coding the facial expression. Classification step is based on SVM method (Support Vector Machine). Experimental results demonstrate that the proposed approach is an effective method to recognize emotions through facial expression with an emotion recognition rate more than90% in real time. This approach is used to control music player based on the variation of the emotional state of the computer user.
Conference Paper
In this paper, we propose a hybrid architecture combining radial basis function network (RBFN) and Principal Component Analysis (PCA) re-constructure model to perform facial expression recognition from static images. The resultant framework is a two stages coarse to fine discrimination model based on local features extracted from eyes and face images by applying PCA technique . It decomposes the acquired data into a small set of characteristic features. The objective of this research is to develop a more efficient approach to classify between seven prototypic facial expressions, such as neutral, joy, anger, surprise, fear, disgust, and sadness. A constructive procedure is detailed and the system performance is evaluated on a public database ”Japanese Females Facial Expression (JAFFE)”. As anticipated, the experimental results demonstrate the potential capabilities of the proposed approach.