Deep Appearance Models
Deep Appearance Models is a method for taking multi-view capture data and creating photorealistic faces that are driveable from cameras mounted on a VR headset. The method consists of two parts. The first is a variational autoencoder to output mesh vertex positions and a 1,024 x 1,024 resolution texture map represented by a small latent code that describes the expression and state of the face. The second part is a method to correspond multi-view capture data and images captured from cameras mounted on a virtual reality headset using a domain adapting variational autoencoder. We combined these two pieces into a real-time system that enables realistic face-to-face conversations in VR.
Example Results
Bibtex
@article{Lombardi:2018,
author = {Lombardi, Stephen and Saragih, Jason and Simon, Tomas and Sheikh, Yaser},
title = {Deep Appearance Models for Face Rendering},
journal = {ACM Trans. Graph.},
issue_date = {August 2018},
volume = {37},
number = {4},
month = jul,
year = {2018},
issn = {0730-0301},
pages = {68:1--68:13},
articleno = {68},
numpages = {13},
publisher = {ACM},
address = {New York, NY, USA},
}
Related Papers
Check out our follow-up work here

