ABSTRACT
We present a new system designed for enabling direct face-to-face interaction for users wearing a head-mounted displays (HMD) in virtual reality environment. Due to HMD's occlusion of a user's face, VR applications and games are mainly designed for single user. Even in some multi-player games, players can only communicate with each other using audio input devices or controllers. To address this problem, we develop a novel system that allows users to interact with each other using facial expressions in real-time. Our system consists of two major components: an automatic tracking and segmenting face processing component and a facial expressions recognizing component based on convolutional neural networks (CNN). First, our system tracks a specific marker on the front surface of the HMD and then uses the extracted spatial data to estimate face positions and rotations for mouth segmentation. At last, with the help of an adaptive approach for histogram based mouth segmentation [Panning et al. 2009], our system passes the processed lips pixels' information to CNN and get the facial expressions results in real-time. The results of our experiments show that our system can effectively recognize the basic expressions of users.
- Burkert, P. Trier, F. Afzal, M. Dengel, A. Liwicki, M, 2015. DeXpression: Deep convolutional neural network for expression recognition. arXiv preprint, pp.1--8.Google Scholar
- Cootes, Timothy, F. and Taylor, C.J., 2004. Statistical Models of Appearance for Computer VisionGoogle Scholar
- Ekman, P. and Wallace V. Friesen, 1975. Pictures of facial affect. consulting psychologists press.Google Scholar
- Krizhevsky, A., Sutskever, I. and Hinton, G.E., 2012. ImageNet Classification with Deep Convolutional Neural Networks. Advances In Neural Information Processing Systems, pp.1--9. Google ScholarDigital Library
- Le Cun, Y. Jackel, l. d. Boser, b. Denker, J. S. Graf, H. P. Guyon, I. Henderson, D. Howard, R. E. Hubbard, W., 1989. Handwritten digit recognition: applications of neural network chips and automatic learning. IEEE Communications Magazine, 27(11), pp.41--46. Google ScholarDigital Library
- Li, H. Trutoiu, L. Olszewski, K. Wei, L. Trutna, T. Hsieh, P. Nicolls, A. Ma, C., 2015. Facial performance sensing head-mounted display. ACM Transactions on Graphics, 34(4), p.47:1--47:9. Google ScholarDigital Library
- Mamic, G., Fookes, C. and Sridharan, S., 2006. What is the average human face? Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 4319 LNCS, pp.692--701. Google ScholarDigital Library
- Marquardt BeautyAnalysis, Inc. The Facial Masks http://www.beautyanalysis.com/research/perfect-face/facial-masks/Google Scholar
- Panning, A. Niese, R. Al-hamadi, A. Michaelis, B., 2009. A new Adaptive Approach for Histogram based Mouth Segmentation. Engineering and Technology, 3(8), pp.779--784.Google Scholar
- Saeed, U. and Dugelay, J.L., 2010. Combining edge detection and region segmentation for lip contour extraction. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 6169 LNCS, pp.11--20. Google ScholarDigital Library
- Tian, Y.T.Y., 2004. Evaluation of Face Resolution for Expression Analysis. 2004 Conference on Computer Vision and Pattern Recognition Workshop, pp.0--6. Google ScholarDigital Library
- Werda, S., Mahdi, W. and Hamadou, A. Ben, 2007. Colour and geometric based model for lip localisation: Application for lip-reading system. Proceedings - 14th International conference on Image Analysis and Processing, ICIAP 2007, (Iciap), pp.9--14. Google ScholarDigital Library
- Zhai, Y. and Shah, M., 2006. Visual Attention Detection in Video Sequences Using Spatiotemporal Cues Categories and Subject Descriptors. Proceedings of the 14th annual ACM international conference on Multimedia, 32816, pp.815--824. Google ScholarDigital Library
Index Terms
- Facial expressions recognition based on convolutional neural networks for mobile virtual reality
Recommendations
Real-time facial tracking in virtual reality
SA '16: SIGGRAPH ASIA 2016 VR ShowcaseVirtual reality (VR) emerges as the next social computing platform. For realizing immersive social interactions, projecting facial expressions onto the virtual avatar a crucial component. This is a challenge in VR as it requires capturing the facial ...
Facial expression recognition with Convolutional Neural Networks
Facial expression recognition has been an active research area in the past 10 years, with growing application areas including avatar animation, neuromarketing and sociable robots. The recognition of facial expressions is not an easy problem for machine ...
Emotion recognition using facial expressions in an immersive virtual reality application
AbstractFacial expression recognition (FER) is an important method to study and distinguish human emotions. In the virtual reality (VR) context, people’s emotions are instantly and naturally triggered and mobilized due to the high immersion and realism of ...
Comments