Video face replacement

We present a method for replacing facial performances in video. Our approach accounts for differences in identity, visual appearance, speech, and timing between source and target videos. Unlike prior work, it does not require substantial manual operation or complex acquisition hardware, only single-...

Full description

Bibliographic Details
Main Authors: Dale, Kevin (Author), Sunkavalli, Kalyan (Author), Johnson, Micah K. (Contributor), Vlasic, Daniel (Author), Matusik, Wojciech (Contributor), Pfister, Hanspeter (Author)
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory (Contributor), Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor)
Format: Article
Language:English
Published: Association for Computing Machinery (ACM), 2012-09-27T15:34:11Z.
Subjects:
Online Access:Get fulltext
Description
Summary:We present a method for replacing facial performances in video. Our approach accounts for differences in identity, visual appearance, speech, and timing between source and target videos. Unlike prior work, it does not require substantial manual operation or complex acquisition hardware, only single-camera video. We use a 3D multilinear model to track the facial performance in both videos. Using the corresponding 3D geometry, we warp the source to the target face and retime the source to match the target performance. We then compute an optimal seam through the video volume that maintains temporal consistency in the final composite. We showcase the use of our method on a variety of examples and present the result of a user study that suggests our results are difficult to distinguish from real video footage.
National Science Foundation (U.S.) (Grant PHY-0835713)
National Science Foundation (U.S.) (Grant DMS-0739255)