Professional Documents
Culture Documents
Associate Professor, School of Electronic Engineering, VIT University, Chennai Campus, Tamil Nadu, India 2
ABSTRACT: In Recent years, the research interest to automated analysis of human affective behaviour has attracted
increasing attention in psychology, bio-medical, human computer interaction, linguistics, neuroscience, computer
science and related disciplines. Since emotions plays an important role in the daily life of human being, the need and
importance of automatic emotion recognition has grown with increasing role of human computer interaction
applications. Emotion places major role of brain cognition for medically monitoring patients emotional states or stress
levels, creating Animated movies, monitoring automobile drivers alertness / drowsiness levels, Human-Robot
interaction, Advanced Gaming which enabling very young children to interact with computers, designing techniques
for forensic identification, etc. This paper survey from the published work papers, since 2013 till date. The paper
surveys about the six prototype expression in human through the various modes of extracting emotion. The paper
detailed about that the face detection and tracking, features extraction mechanisms and the technique used for
expression classification by the face modelling of state-of-the-art classification. The paper also discussed on
multimodal expression classifier of human, i.e. from body and face. Notes also presented the tracking human through
sensor and ends with the challenges and future with intention of helping students and researchers who are new to this
field.
KEYWORDS: Modes of extracting emotion, expression recognition, face modelling, state-of-the-art, basic six
emotion feature tracking, body tracking, and Microsoft Kinect sensor.
I.INTRODUCTION
th
In early Aristotelian era (4 century BC) study on the Physiognomy and Facial Expression. [1] Physiognomy is the
assessment of a person personality or character from their outer appearance, especially face. But over the year,
Physiognomy has been not interested, but facial expression continuously been active topic. From the foundational
studies that formed the basis of todays research on facial expression can be tracked to 17 th century. In 1649, John
Bulwer in his book Pathomyotmia was given details of muscle movement of head and various expressions in human.
[2] In 1667, Le Bruns lectures the physiognomy; later reproduce the book in 1734 [3]. In 18th century actors and artist
referred to his book to achieve the perfect imitation of genuine facial expression. In 19 th century, a facial expression
analysis has direct relationship to the automatic face analysis and face expression recognition this was brought by
Darwin. [4] In 1894, Darwin states that principles of expression in both humans and animals also groped the various
kind of expression and cataloged the facial deformation. [5] In 1884, William James states James Lange theory the
all emotion are derived from the presence of stimulus in body which evokes the physiological response.
[6]Another important study of facial expression analysis and human emotion is done by the Paul Ekman in 1967. He
developed the facial expression recognizer, and analyst the facial expression muscular movement showed that different
emotion in human through the photographic stimuli. In 1969, [7]Wallbott and Scherer determined emotion from the
body movement and speech. In 1978, Suwa established the automatic recognition of facial expression system for
analysis the facial expression from a sequence of image of tacking points. Although this system might not be clearly
seen the tracking points till 1990s. Later 1992, Samal and Iyengar presented the Facial Feature and Expression analysis
of tacking points in movie frames and they also states that the automatic recognition of facial expression requires robust
face detection and face tracking system Early and late of 1990, led to the development of robotic face detection, face
tracking, face expression analysis and at the same time become very popularity of fields are the Human computer
Interaction (HCI), Affective Computing, etc.,. Since 1990s, researchers are started to develop of interest on automatic
facial expression recognition become very active. From 1990 to till now, researchers are mostly concentrating on
automatic facial expression recognition and emotion in human using various modes of extraction
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1966
NON-PHYSIOLOGICAL SIGNAL:
1.1. Speech signal:1.Tin Lay New, Say Wei Foo, Liyanage C. De Silva, Speech emotion recognition using hidden
Markov models, 2003 [10] , proposes an emotion recognition from the speech signal using methods of short time log
frequency power coefficients(LFPC) to represent the speech signals and a discrete hidden Markov model (HMMs). The
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1967
10.15662/ijareeie.2015.0404017
1968
10.15662/ijareeie.2015.0404017
1969
10.15662/ijareeie.2015.0404017
1970
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1971
10.15662/ijareeie.2015.0404017
1972
10.15662/ijareeie.2015.0404017
1973
10.15662/ijareeie.2015.0404017
1974
TECHNIQUE USED
EXPERIMENTAL RESULT
[10]
1.
2.
[11]
1.
[12]
[13]
2.
3.
4.
1.
2.
Seven cues:
symmetry and limb alignment, head alignment and
offset, body openness, average rate of change, relative
movement, smooth-jerk, and location of activity
2.
[14]
1.
2.
3.
[15]
1.
2.
3.
Copyright to IJAREEIE
rate
10.15662/ijareeie.2015.0404017
4.
5.
6.
[16]
1.
2.
3.
[17]
1.
2.
[18]
1.
2.
[20,21,22]
1.
2.
3.
4.
1.
1.
2.
3.
4.
Template matching
Segmentation
Rubber snake model
Color snake model
1.
[27,28,29,30]
1.
2.
3.
4.
1.
2.
3.
[31,32,33]
1.
2.
3.
1.
2.
[34,35]
1.
2.
3.
Viterbi algorithm
Histogram of Oriented Gradients descriptor
Multiple linear regression models
1.
Expression
recognition
is
possible when using active
appearance model
Part based model has good
accuracy
Head pose variation is possible
[23,24,25,26]
2.
3.
2.
2.
3.
[36,37]
1.
2.
3.
1.
2.
3.
[38,39]
1.
2.
3.
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1976
[41]
1.
2.
3.
[42,43]
[44] -- [52]
1.
2.
3.
4.
5.
6.
7.
Direct Parameterizations
Face Action coding system (FACS)
MPEG4 Facial Animation Parameters (FAP)
Abstract Muscle Actions (AMA)
Minimal Perception Action
Example-based deformation
mass-spring network model
[53]
1.
2.
3.
Eigen faces
Eigen vector
Principle Component Analysis (PCA)
1.
2.
[54]
3.
1.
2.
3.
4.
5.
1.
1.
2.
3.
4.
5.
1.
2.
3.
4.
Morphing model
Segmentation
Automatic matching algorithm
Optical flow algorithm
Illumination-Corrected Texture Extraction
KanadeLucasTomasi (KLT) tracker
Eigen faces
Active Shape model
Constrained Local appearance (CLM)
[40]
1.
2.
[55]
[56,57,31]
1.
2.
[58]
1.
2.
3.
4.
Copyright to IJAREEIE
2.
3.
[44] -- [48]
1.
2.
3.
4.
1.
2.
3.
10.15662/ijareeie.2015.0404017
1977
4.
[59]
[60]
1.
2.
3.
4.
5.
1.
1.
2.
3.
1.
2.
4.
2.
Discussion: The comparison table gives the various techniques used for estimate the face expression recognition
observed in all references. From the [10] - [18] literature survey the facial expression system achieved from their
various modes of extracting emotion. In this, face gesture has achieved more advantages such as high accuracy of
emotion, head pose variation, very robustly compared to the other modes of extracting. From [20]- [57] reviewed that
defines the automatic facial analysis and state-of-art model. From [20]- [39] defines face analysis model, retrieved such
as a feature points, feature detection of face, shape and texture model, deformable model, energy function, face
expression recognizer, head pose of face feature. From [40]- [52] reviewed that defines the facial animation parameter,
retrieved such as morphable 3d face model, animation facial expression recognition also achieved the face expression
but accuracy is low, no head pose variation, etc.,
From [53]- [57] and [44]- [48] defines the facial analysis by facial animation parameter for facial expression
recognition system and head pose variation. Its achieved high accuracy of face expression and emotion recognition,
head pose variations, very robustly (particularly these two models are Candide model and two methods of Active
Appearance Model (AAM)) compared to the other face model In [58]- [61] reviewed the estimates a body tracking for
emotion recognition using various sensor, particularly Microsoft Kinect sensor [62] has achieved high accuracy, very
robustly skeletal tracking algorithm, also has a face expression classifier and head pose variation, compare to the
various sensors. From these, we reviewed that the discussion and comparison of technique for automatic expression
recognition.
VII.FUTURE WORK AND CHALLENGES
As discussed earlier, the major challenge that researchers are focusing on capturing spontaneous expression on images
and video to get high accuracy, avoid the loss of authenticity, and also spontaneous head pose expression variation.
Another challenging capturing spontaneous non basic expression compared to the capturing of basic expression and
also achieves the same accuracy of basic six emotions. Another area, where more work is carried out on completely
automatic expression system in real time application and expression synthesis. Researchers have also focus for
multimodal fusion of expression recognition has achieved less accuracy when compared to the unimodal system. Also
differences do exist in facial feature and facial expression between cultures, age groups is more challenging. In medical
application, some patient likes autistic disorder, Asperger syndrome, epilepsy, etc. monitoring for the loss of facial
expression analysis also challenging.
In skeletal body tracking, loss their authenticity when expression analysis is also challenging. I conclude here, these
are future work and challenging of expression analysis
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1978
Richard Wiseman and Rob Jenkins Roger Highfield, "How your looks betray your personality".: New Scientist, Feb 2009.
[Online]. http://www.library.northwestern.edu/spec/hogarth/physiognomics1.html.
[3]
J Montagu., "The Expression of the Passions: The Origin and Influence of Charles Le Bruns Confrence sur l'expression gnrale et particulire,," Yale
University press, 1994.
[4]
Charles Darwin, The Expression of the Emotions in Man and Animals,, 2nd ed., J Murray, Ed. London , 1904.
[5]
William James, "What is Emotion," Mind, vol. 9, no. 34, pp. 188-205, 1884.
[6]
Paul Ekman and Wallace.V. Friesen, "Pan-Cultural Elements In Facial Display Of Emotion," Science, pp. 86-88, 1969.
[7]
Harald G., and Klaus R. Scherer Wallbott, ""Cues and channels in emotion recognition."," Journal of personality and social psychology, pp. 69-70, 1986.
[8]
F., Malfaz, M., Sequeira, J., Gorostiza, J. F., & Salichs, M. A. Alonso-Martn, "A Multimodal Emotion Detection System during Human-Robot Interaction,"
Sensors, vol. 13(11), no. 15549., 2013.
[9]
Stefanos Kollias and Kostas Karpouzis, "Multimodal Emotion Recognition And Expressivity Analysis," IEEE International Conference on Mulitimedia and
Expo, pp. 779-783, 2005.
[10]
Say Wei Foo,LiyanageC. De Silva Tin Lay New, "Speech emotion recognition using hidden Markov models," Speech communication , no. 41.4, pp. 603-623,
2003.
[11]
K. Takahashi and R. Nakatsu Nicholson, "Emotion Recognition in Speech Using Neural Networks," Neural Computing & Applications , no. 9.4, pp. 290-296,
2000.
[12]
Ze-Jing Chuang and Chung-Hsien Wu, "Multi-Modal Emotion Recognition from Speech and Text," International Journal of Computational Linguistics and
Chinese Language Processing, no. 9.2 , pp. 1-18, 2004..
[13]
Michael, and Mei Si, Garber Barron, "Using body movement and posture for emotion detection in non-acted scenarios. ," , 2012.
[14]
I. Kotsia and I. Pitas, "Facial Expression Recognition in Image Sequences Using Geometric Deformation Features and Support Vector Machines," IEEE
Trans. Image Processing, vol. 16, no. 1, pp. 172-187, 2007.
[15]
George Caridakis, "Multimodal emotion recognition from expressive faces, body gestures and speech.", Artificial intelligence and innovations 2007: From
theory to applications Springer US, pp. 375-388, 2007.
[16]
Peter C. W. Fung, Siew E. Chua and Grainne M. McAlonan Teresa K. W. Wong, "Abnormal spatiotemporal processing of emotional facial expressions in
childhood autism: dipole source analysis of eventrelated potentials," European Journal of Neuroscience ():, no. 28.2, pp. 407-416., 2008.
[17]
Panagiotis C. Petrantonakis and Leontios J. Hadjileontiadis, "Emotion Recognition from Brain Signals Using Hybrid Adaptive Filtering and Higher Order
Crossings Analysis," 2010 Affective Computing, IEEE Transactions , no. 1.2, pp. 81-97, 2010.
[18]
Marini, and Abdul Wahab Othman, "Affective face processing analysis in autism using electroencephalogram.," in Information and Communication
Technology for the Muslim World (ICT4M), 2010 International Conference on. IEEE, 2010.
[19]
Hanan Salam, "Multi-Object modeling of the face," Universit Rennes 1, Signal and Image processing, Thesis 2013.
[20]
Peter W.Hallinan and David S.Cohen Alan L Yulie, "Feature extraction from faces using deformable templates", International journal of computer vision, vol.
8(2), pp. 99111, 1992.
[21]
Zhang, and Ruan Qiuqi Baizhen, "Facial feature extraction using improved deformable templates.," in Signal Processing, 2006 8th International Conference
on,Vol 4, 2006.
[22]
Marius, and Francoise J. Preteux. Malciu, "Tracking facial features in video sequences using a deformable-model-based approach," in International
Symposium on Optical Science and Technology. International Society for Optics and Photonics, 2000.
[23]
A. Witkin, and D. Terzopoulos M. Kass, "Snakes: Active contour models," International journal of computer vision, vol. 1(4), pp. 321-331, 1988.
[24]
Petia, and Eric Mart Radeva, "Facial features segmentation by model-based snakes," in International Conference on Computing Analysis and Image
Processing, Prague, 1995.
[25]
Won kim,Changmok Oh and Ju-Jang Lee Kap-ho Seo, "Face detection and facial feature extraction using color snake," in Industrial Electronics, 2002. ISIE
2002. Proceedings of the 2002 IEEE International Symposium, 2002.
[26]
Rein-Lien, and Anil K. Jain Hsu, "Generating discriminating cartoon faces using interacting snakes," Pattern Analysis and Machine Intelligence, IEEE
Transactions, vol. 25, no. 11, pp. 1388-1398, 2003.
[27]
D.H. Cooper, C.J. Taylor, and J. Graha, T.F. Cootes, "Trainable method of parametric shape description," Image and Vision Computing, vol. 10, no. 5, pp.
289294, 1992.
[28]
C.J. Taylor, D.H. Cooper, and J. Graham T.F. Cootes, "Active shape models-their training and application," Computer vision and image understanding, vol.
61, no. 1, pp. 3859, 1995.
[29]
Yuanzhong, and Wataru Ito Li, "Shape parameter optimization for adaboosted active shape model.", in Computer Vision, 2005. ICCV 2005. Tenth IEEE
International Conference on vol 1, 2005.
[30]
David, and Timothy F. Cootes Cristinacce, "Boosted Regression Active Shape Models," British Machine Vision Conference, 2007.
[31]
David, and Timothy F. Cootes Cristinacce, "Feature Detection and Tracking with Constrained Local Models," British Machine Vision Conference, vol. 2, no.
5, 2006.
[32]
David, and Timothy F. Cootes Cristinacce, "Facial feature detection and tracking with automatic template selection.," in Automatic Face and Gesture
Recognition, 2006. FGR 2006. 7th International Conference on. IEEE, 2006.
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1979
Copyright to IJAREEIE
10.15662/ijareeie.2015.0404017
1980