• DocumentCode
    3672292
  • Title

    Joint patch and multi-label learning for facial action unit detection

  • Author

    Kaili Zhao; Wen-Sheng Chu;Fernando De la Torre;Jeffrey F. Cohn;Honggang Zhang

  • Author_Institution
    School of Comm. and Info. Engineering, Beijing University of Posts and Telecom., China
  • fYear
    2015
  • fDate
    6/1/2015 12:00:00 AM
  • Firstpage
    2207
  • Lastpage
    2216
  • Abstract
    The face is one of the most powerful channel of nonverbal communication. The most commonly used taxonomy to describe facial behaviour is the Facial Action Coding System (FACS). FACS segments the visible effects of facial muscle activation into 30+ action units (AUs). AUs, which may occur alone and in thousands of combinations, can describe nearly all-possible facial expressions. Most existing methods for automatic AU detection treat the problem using one-vs-all classifiers and fail to exploit dependencies among AU and facial features. We introduce joint-patch and multi-label learning (JPML) to address these issues. JPML leverages group sparsity by selecting a sparse subset of facial patches while learning a multi-label classifier. In four of five comparisons on three diverse datasets, CK+, GFT, and BP4D, JPML produced the highest average F1 scores in comparison with state-of-the art.
  • Keywords
    "Gold","Correlation","Face","Joints","Feature extraction","Encoding","Convergence"
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision and Pattern Recognition (CVPR), 2015 IEEE Conference on
  • Electronic_ISBN
    1063-6919
  • Type

    conf

  • DOI
    10.1109/CVPR.2015.7298833
  • Filename
    7298833