A new framework has been created that can interpret mouth gestures as a medium for interaction within virtual reality environment in real-time.

‘Mouth gestures can be interpreted within virtual reality environment in real-time with the help of the new technology developed.’

The study was conducted by the research team at Binghamton University, State University of New York.
To combat this issue, Binghamton University Professor of Computer Science Lijun Yin and his team created a new framework that interprets mouth gestures as a medium for interaction within virtual reality in real-time.




Yin's team tested the application on a group of graduate students. Once a user put on a head-mounted display, they were presented with a simplistic game; the objective of the game was to guide the player's avatar around a forest and eat as many cakes as possible.
Players had to select their movement direction using head rotation, move using mouth gestures and could only eat the cake by smiling. The system was able to describe and classify the user's mouth movements, and it achieved high correct recognition rates. The system has also been demonstrated and validated through a real-time virtual reality application.
"We hope to make this applicable to more than one person, maybe two. Think Skype interviews and communication," said Yin.
"Imagine if it felt like you were in the same geometric space, face to face, and the computer program can efficiently depict your facial expressions and replicate them so it looks real."
Advertisement
"The virtual world isn't only for entertainment. For instance, health care uses VR to help disabled patients," said Yin.
Advertisement
Students Umur Aybars Ciftci and Xing Zhang contributed to this research.
The paper, "Partially occluded facial action recognition and interaction in virtual reality applications," was presented at the 2017 IEEE International Conference on Multimedia and Expo.
Source-Eurekalert