EXPERTNet
2018, Proceedings of the 11th Indian Conference on Computer Vision, Graphics and Image Processing
Abstract
Facial expressions have essential cues to infer the humans state of mind, that conveys adequate information to understand individuals' actual feelings. Thus, automatic facial expression recognition is an interesting and crucial task to interpret the humans cognitive state through the machine. In this paper, we proposed an Exigent Features Preservative Network (EXPERTNet), to describe the features of the facial expressions. The EXPERTNet extracts only pertinent features and neglect others by using exigent feature (ExFeat) block, mainly comprises of elective layer. Specifically, elective layer selects the desired edge variation features from the previous layer outcomes, which are generated by applying different sized filters as 1 1, 3 3, 5 5 and 7 7. Different sized filters aid to elicits both micro and high-level features that enhance the learnability of neurons. ExFeat block preserves the spatial structural information of the facial expression, which allows to discriminate between different classes of facial expressions. Visual representation of the proposed method over different facial expressions shows the learning capability of the neurons of different layers. Experimental and comparative analysis results over four comprehensive datasets: CK+, MMI DISFA and GEMEP-FERA, ensures the better performance of the proposed network as compared to existing networks.
References (30)
- P. Ekman, (1993). Facial expression and emotion. American Psychologist, 48(4), 384.
- P. Ekman and W.V. Friesen, (1977). Facial action coding system.
- W. V. Friesen and P. Ekman, (1983). EMFACS-7: Emotional facial action coding system. Unpublished manuscript, University of California at San Francisco, 2(36), 1.
- I. A. Essa and A. P. Pentland, (1997). Coding, analysis, interpretation, and recognition of facial expressions. IEEE transactions on pattern analysis and machine intelligence, 19(7), 757-763.
- I. Kotsia and I. Pitas, (2007). Facial expression recognition in image sequences using geometric deformation features and support vector machines. IEEE transactions on image processing, 16(1), 172-187.
- D. Gabor, (1946). Theory of communication. Part 1: The analysis of information. Journal of the Institution of Electrical Engineers-Part III: Radio and Communication Engineering, 93(26), 429-441.
- C. Shan, S. Gong and P.W. McOwan (2009). Facial expression recognition based on local binary patterns: A comprehensive study. Image and vision Computing, 27(6), 803-816.
- C. C. Lai and C. H. Ko, (2014). Facial expression recognition based on two-stage features extraction. Optik-International Journal for Light and Electron Optics, 125(22), 6678-6680.
- T. Jabid, M. H. Kabir and O. Chae, (2010). Robust facial expression recognition based on local directional pattern. ETRI journal, 32(5), 784-794.
- A. R. Rivera, J. R. Castillo and O. O. Chae, (2013). Local directional number pattern for face analysis: Face and expression recognition. IEEE transactions on image processing, 22(5), 1740- 1752.
- A. R. Rivera, J. R. Castillo and O. Chae (2015). Local directional texture pattern image descriptor. Pattern Recognition Letters, 51, 94- 100.
- B. Ryu, A. R. Rivera, J. Kim and O. Chae, (2017). Local directional ternary pattern for facial expression recognition. IEEE Transactions on Image Processing, 26(12), 6006-6018.
- A. Krizhevsky, I. Sutskever and G. E Hinton, (2012). Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, 1097-1105.
- K. Simonyan and A. Zisserman (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
- C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens and Z. Wojna, (2016). Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2818-2826.
- K. He, X. Zhang, S. Ren and J. Sun (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778.
- A. Mollahosseini, D. Chan and M. H. Mahoor, (2016). Going deeper in facial expression recognition using deep neural networks. In Applications of Computer Vision (WACV), 1-10.
- B. Hasani and M. H. Mahoor, (2017). Spatio-temporal facial expression recognition using convolutional neural networks and conditional random fields. In 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), 790-795.
- H. Jung, S. Lee, S. Park, I. Lee, C. Ahn and J. Kim, (2015). Deep temporal appearance-geometry network for facial expression recognition. arXiv preprint arXiv:1503.01532.
- P. Khorrami, T. Paine, and T. Huang, (2015). Do deep neural networks learn facial action units when doing expression recognition?. In Proceedings of the IEEE International Conference on Computer Vision Workshops, 19-27.
- H. Ding, S. K. Zhou and R. Chellappa, (2017). Facenet2expnet: Regularizing a deep face recognition net for expression recognition. In 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), 118-126.
- Y. Kim, B. Yoo, Y. Kwak, C. Choi and J. Kim, (2017). Deep generative-contrastive networks for facial expression recognition. arXiv preprint arXiv:1703.07140.
- P. Burkert, F. Trier, M. Z. Afzal, A. Dengel and M. Liwicki, (2015). Dexpression: Deep convolutional neural network for expression recognition. arXiv preprint arXiv:1509.05371.
- K. Zhang, Y. Huang, Y. Du and L. Wang, (2017). Facial expression recognition based on deep evolutional spatial-temporal networks. IEEE Transactions on Image Processing, 26(9), 4193- 4203.
- P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar and I. Matthews, (2010). The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 94-101).
- M. Pantic, M. Valstar, R. Rademaker and L. Maat, (2005,). Web- based database for facial expression analysis. In IEEE international conference on multimedia and Expo, 5.
- M. Valstar and M. Pantic, (2010). Induced disgust, happiness and surprise: an addition to the mmi facial expression database. In Proc. 3rd Intern. Workshop on EMOTION (satellite of LREC): Corpora for Research on Emotion and Affect, 65.
- S. M. Mavadati, M. H. Mahoor, K. Bartlett, P. Trinh and J. F. Cohn, (2013). Disfa: A spontaneous facial action intensity database. IEEE Transactions on Affective Computing, 4(2), 151-160.
- M. F. Valstar, B. Jiang, M. Mehu, M. Pantic and K. Scherer, (2011). The first facial expression recognition and analysis challenge. In IEEE International Conference on Automatic Face & Gesture Recognition and Workshops (FG 2011), 921-926.
- P.Viola, and M. J. Jones, (2004). Robust real-time face detection. International journal of computer vision, 57(2), 137-154.