FaceVR: Real-Time Facial Reenactment and Eye Gaze Control in VR

November 25, 2016

Justus Thies et al., have developed a novel approach for real-time gaze-aware facial capture system to drive the photo-realistic reconstructed digital face in virtual reality. Their approach enables facial reenactment that can transfer facial expressions and realistic eye appearance between a source and a target actor video.

Real-Time Facial Reenactment and Eye Gaze Control in VR
Source: [FaceVR: Real-Time Facial Reenactment and Eye Gaze Control in VR](https://arxiv.org/abs/1610.03151)

Head mounted displays (HMDs) provide immersive renderings of virtual environments. But in order to do so HMDs block the majority of the actor/participant’s face. In order to reconstruct the actor/participant’s face, Thies et al., use an RGB camera to capture the facial performance of the participant and Pupil Labs’s Oculus Rift DK2 add-on cup to capture eye movements within the HMD. The source actor’s facial and eye movement data is then used to drive the photo-realistic facial animations of the target video, therefore enabling gaze-aware facial reenactment.

Check out their full paper available on arxiv - FaceVR: Real-Time Facial Reenactment and Eye Gaze Control in VR.

If you use Pupil in your research and have published work, please send us a note. We would love to feature the work here on the blog and in a list of work that cites Pupil.