Facial asymmetry quantification for expression invariant human identification

Yanxi Liu, Karen L. Schmidt, Jeffrey F. Cohn, Sinjini Mitra

Research output: Contribution to journalArticlepeer-review

108 Scopus citations


We investigate facial asymmetry as a biometric under expression variation. For the first time, we have defined two types of quantified facial asymmetry measures that are easily computable from facial images and videos. Our findings show that the asymmetry measures of automatically selected facial regions capture individual differences that are relatively stable to facial expression variations. More importantly, a synergy is achieved by combining facial asymmetry information with conventional EigenFace and FisherFace methods. We have assessed the generality of these findings across two publicly available face databases: Using a random subset of 110 subjects from the FERET database, a 38% classification error reduction rate is obtained. Error reduction rates of 45-100% are achieved on 55 subjects from the Cohn-Kanade AU-Coded Facial Expression Database. These results suggest that facial asymmetry may provide complementary discriminative information to human identification methods, which has been missing in automatic human identification.

Original languageEnglish (US)
Pages (from-to)138-159
Number of pages22
JournalComputer Vision and Image Understanding
Issue number1-2
StatePublished - 2003

All Science Journal Classification (ASJC) codes

  • Software
  • Signal Processing
  • Computer Vision and Pattern Recognition


Dive into the research topics of 'Facial asymmetry quantification for expression invariant human identification'. Together they form a unique fingerprint.

Cite this