TY - JOUR
T1 - High resolution acquisition, learning and transfer of dynamic 3-D facial expressions
AU - Wang, Yang
AU - Huang, Xiaolei
AU - Lee, Chan Su
AU - Zhang, Song
AU - Li, Zhiguo
AU - Samaras, Dimitris
AU - Metaxas, Dimitris
AU - Elgammal, Ahmed
AU - Huang, Peisen
PY - 2004
Y1 - 2004
N2 - Synthesis and re-targeting of facial expressions is central to facial animation and often involves significant manual work in order to achieve realistic expressions, due to the difficulty of capturing high quality dynamic expression data. In this paper we address fundamental issues regarding the use of high quality dense 3-D data samples undergoing motions at video speeds, e.g. human facial expressions. In order to utilize such data for motion analysis and re-targeting, correspondences must be established between data in different frames of the same faces as well as between different faces. We present a data driven approach that consists of four parts: 1) High speed, high accuracy capture of moving faces without the use of markers, 2) Very precise tracking of facial motion using a multi-resolution deformable mesh, 3) A unified low dimensional mapping of dynamic facial motion that can separate expression style, and 4) Synthesis of novel expressions as a combination of expression styles. The accuracy and resolution of our method allows us to capture and track subtle expression details. The low dimensional representation of motion data in a unified embedding for all the subjects in the database allows for learning the most discriminating characteristics of each individual's expressions as that person's "expression style". Thus new expressions can be synthesized, either as dynamic morphing between individuals, or as expression transfer from a source face to a target face, as demonstrated in a series of experiments.
AB - Synthesis and re-targeting of facial expressions is central to facial animation and often involves significant manual work in order to achieve realistic expressions, due to the difficulty of capturing high quality dynamic expression data. In this paper we address fundamental issues regarding the use of high quality dense 3-D data samples undergoing motions at video speeds, e.g. human facial expressions. In order to utilize such data for motion analysis and re-targeting, correspondences must be established between data in different frames of the same faces as well as between different faces. We present a data driven approach that consists of four parts: 1) High speed, high accuracy capture of moving faces without the use of markers, 2) Very precise tracking of facial motion using a multi-resolution deformable mesh, 3) A unified low dimensional mapping of dynamic facial motion that can separate expression style, and 4) Synthesis of novel expressions as a combination of expression styles. The accuracy and resolution of our method allows us to capture and track subtle expression details. The low dimensional representation of motion data in a unified embedding for all the subjects in the database allows for learning the most discriminating characteristics of each individual's expressions as that person's "expression style". Thus new expressions can be synthesized, either as dynamic morphing between individuals, or as expression transfer from a source face to a target face, as demonstrated in a series of experiments.
UR - http://www.scopus.com/inward/record.url?scp=4644229847&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=4644229847&partnerID=8YFLogxK
U2 - 10.1111/j.1467-8659.2004.00800.x
DO - 10.1111/j.1467-8659.2004.00800.x
M3 - Article
AN - SCOPUS:4644229847
SN - 0167-7055
VL - 23
SP - 677
EP - 686
JO - Computer Graphics Forum
JF - Computer Graphics Forum
IS - 3 SPEC. ISS.
ER -