Browse > Article
http://dx.doi.org/10.3837/tiis.2016.03.026

A Local Feature-Based Robust Approach for Facial Expression Recognition from Depth Video  

Uddin, Md. Zia (Department of Computer Education, Sungkyunkwan University)
Kim, Jaehyoun (Department of Computer Education, Sungkyunkwan University)
Publication Information
KSII Transactions on Internet and Information Systems (TIIS) / v.10, no.3, 2016 , pp. 1390-1403 More about this Journal
Abstract
Facial expression recognition (FER) plays a very significant role in computer vision, pattern recognition, and image processing applications such as human computer interaction as it provides sufficient information about emotions of people. For video-based facial expression recognition, depth cameras can be better candidates over RGB cameras as a person's face cannot be easily recognized from distance-based depth videos hence depth cameras also resolve some privacy issues that can arise using RGB faces. A good FER system is very much reliant on the extraction of robust features as well as recognition engine. In this work, an efficient novel approach is proposed to recognize some facial expressions from time-sequential depth videos. First of all, efficient Local Binary Pattern (LBP) features are obtained from the time-sequential depth faces that are further classified by Generalized Discriminant Analysis (GDA) to make the features more robust and finally, the LBP-GDA features are fed into Hidden Markov Models (HMMs) to train and recognize different facial expressions successfully. The depth information-based proposed facial expression recognition approach is compared to the conventional approaches such as Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Linear Discriminant Analysis (LDA) where the proposed one outperforms others by obtaining better recognition rates.
Keywords
Depth video; local binary patterns (LBP); generalized discriminant analysis (GDA); hidden Markov models (HMMs);
Citations & Related Records
연도 인용수 순위
  • Reference
1 Donato G., Bartlett M. S., Hagar J. C., Ekman P., and Sejnowski T. J., “Classifying Facial Actions,” IEEE Transaction on Pattern Analysis and Machine Intelligence, Vol. 21, No. 10, pp. 974-989, 1999. Article (CrossRef Link)   DOI
2 Dubuisson S., Davoine F., and Masson M., “A solution for facial expression representation and recognition,” Signal Processing: Image Communication, Vol. 17, pp. 657-673, 2002. Article (CrossRef Link)   DOI
3 Chao-Fa C. and Shin F. Y., “Recognizing Facial Action Units Using Independent Component Analysis and Support Vector Machine,” Pattern Recognition, Vol. 39, pp. 1795-1798, 2006. Article (CrossRef Link)   DOI
4 Ojala T., Pietikäinen M., Mäenpää T., “Multiresolution gray scale and rotation invariant texture analysis with local binary patterns,” IEEE Trans. Pattern Anal. Mach. Intell, Vol. 24, pp. 971-987, 2002. Article (CrossRef Link)   DOI
5 Shan C., Gong S., and McOwan P., “Facial expression recognition based on local binary patterns: A comprehensive study,” Image Vis. Comput., Vol. 27, pp. 803-816, 2009. Article (CrossRef Link)   DOI
6 Shen L., Bai L., and Fairhurst M.,” Gabor wavelets and General Discriminant Analysis for face identification and verification,” Image and Vision Computing, Vol. 25, pp. 553-563, 2007. Article (CrossRef Link)   DOI
7 Uddin M. Z. and Hassan M. M., "A Depth Video-Based Facial Expression Recognition System Using Radon Transform, Generalized Discriminant Analysis, and Hidden Markov Model," Multimedia Tools And Applications, Vol. 74, No. 11, pp. 3675-3690, 2015. Article (CrossRef Link)   DOI
8 P. Yu, D. Xu, and P. Yu "Comparison of PCA, LDA and GDA for Palm print Verification," in Proc. of International Conference on Information, Networking and Automation, pp. 148-152, 2010. Article (CrossRef Link)
9 Y. Wang, K. Huang, and T. Tan, "Human activity recognition based on R transform," in Proc. of IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-8, 2007. Article (CrossRef Link)
10 H.S. Koppula, R. Gupta, and A. Saxena, “Human activity learning using object affordances from rgb-d videos,” International Journal of Robotics Research, vol. 32, no. 8, pp. 951-970, 2013. Article (CrossRef Link)   DOI
11 X. Yang and Y. Tian, "Eigenjoints-based action recognition using naive-bayesnearest-neighbor," in Proc. of Workshop on Human Activity Understanding from 3D Data, pp. 14-19, 2012. Article (CrossRef Link)
12 J. Sung, C. Ponce, B. Selman, and A. Saxena, "Unstructured human activity detection from rgbd images," in Proc. of IEEE International Conference on Robotics and Automation, pp. 842-849, 2012. Article (CrossRef Link)
13 A. McCallum, D. Freitag, and F.C.N. Pereira, "Maximum entropy markov models for information extraction and segmentation," in Proc. of International Conference on Machine Learning, pp. 591-598,2000. Article (CrossRef Link)
14 H. Hamer, K. Schindler, E. Koller-Meier, and L. Van Gool, "Tracking a hand manipulating an object," in Proc. of IEEE International Conference on Computer Vision, pp. 1475-1482, 2009. Article (CrossRef Link)
15 I. Oikonomidis, N. Kyriazis, and A.A. Argyros, "Tracking the articulated motion of two strongly interacting hands ," in Proc. of IEEE Conference on Computer Vision and Pattern Recognition, pp. 1862-1869, 2012. Article (CrossRef Link)
16 H. Hamer, J. Gall, T. Weise, and L. Van Gool, "An object-dependent hand pose prior from sparse training data," in Proc. of IEEE Conference on Computer Vision and Pattern Recognition. pp. 671-678, 2010. Article (CrossRef Link)
17 W Li., Z. Zhang, and. Z. Liu, "Action recognition based on a bag of 3d points," in Proc. of Workshop on Human Activity Understanding from 3D Data, pp. 9-14, 2010. Article (CrossRef Link)
18 Rahman M. T. and Kehtarnavaz N., “Real-Time Face-Priority Auto Focus for Digital and Cell-Phone Cameras,” IEEE Transactions on Consumer Electronics , Vol. 54 , No. 4 , pp.1506 -1513 , 2008. Article (CrossRef Link)   DOI
19 Calder A. J., Burton A. M., Miller P., Young A. W., and Akamatsu S., “A principal component analysis of facial expressions,” Vision Research, Vol. 41, pp. 1179-1208, 2001. Article (CrossRef Link)   DOI
20 Yu P., Xu D., and Yu P., "Comparison of PCA, LDA and GDA for Palm print Verification," in Proc. of the International Conference on Information, Networking and Automation, pp.148-152, 2010. Article (CrossRef Link)
21 W Li., Z. Zhang, and. Z. Liu, “Expandable data-driven graphical modeling of human actions based on salient postures,” IEEE Transactions on Circuits and Systems for Video Technology, Vol. 18, No. 11, pp. 1499-1510, 2008. Article (CrossRef Link)   DOI
22 O. Oreifej and Z. Liu, "Hon4d: Histogram of oriented 4d normals for activity recognition from depth sequences," in Proc. of IEEE Conference on Computer Vision and Pattern Recognition, pp. 716-723, 2013. Article (CrossRef Link)
23 A. Vieira, E. Nascimento, G. Oliveira, Z. Liu, and M. Campos, "Stop: Space-time occupancy patterns for 3d action recognition from depth map sequences," in Proc. of Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications, pp. 252-259, 2012. Article (CrossRef Link)
24 J.Wang, Z. Liu, J. Chorowski, Z. Chen, and Y.Wu, " Robust 3d action recognition with random occupancy patterns," in Proc. of European Conference on Computer Vision, pp. 872-885, 2012. Article (CrossRef Link)
25 X. Yang, C. Zhang, and Y. Tian, "Recognizing actions using depth motion mapsbased histograms of oriented gradients," in Proc. of ACM International Conference on Multimedia, pp. 1057-1060, 2012. Article (CrossRef Link)
26 J. Lei, X. Ren, and D. Fox, "Fine-grained kitchen activity recognition using rgb-d," in Proc. of ACM Conference on Ubiquitous Computing, pp.208-211, 2012.Article (CrossRef Link)
27 A. Jalal ., M.Z. Uddin, J.T. Kim, and T.S. Kim, “Recognition of human home activities via depth silhouettes and R transformation for smart homes,” Indoor and Built Environment, vol. 21, no 1, pp. 184-190, 2011. Article (CrossRef Link)   DOI
28 H.D. Yang, S. Sclaroff, and S.W. Lee, “Sign language spotting with a threshold model based on conditional random fields,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no.7, pp.1264-1277, 2009. Article (CrossRef Link)   DOI
29 S. Ong and S. Ranganath, “Automatic sign language analysis: A survey and the future beyond lexical meaning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 6, pp. 873-891, 2005. Article (CrossRef Link)   DOI
30 T. Pei, T. Starner, H. Hamilton, I. Essa, and J. Rehg, "Learnung the basic units in american sign language using discriminative segmental feature selection," in Proc. of IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 4757-4760, 2009. Article (CrossRef Link)
31 J. Bukhari, M. Rehman, S. I. Malik, A. M. Kamboh, and A. Salman, “American Sign Language Translation through Sensory Glove; SignSpeak,” International Journal of u- and eService, Science and Technology, Vol.8, No.1, pp.131-142, 2015. Article (CrossRef Link)   DOI