Joint Patch and Multi-label Learning for Facial Action Unit Detection
Abstract
The face is one of the most powerful channel of nonverbal communication. The most commonly used taxonomy to describe facial behaviour is the Facial Action Coding System (FACS). FACS segments the visible effects of facial muscle activation into 30+ action units (AUs). AUs, which may occur alone and in thousands of combinations, can describe nearly all-possible facial expressions. Most existing methods for automatic AU detection treat the problem using one-vs-all classifiers and fail to exploit dependencies among AU and facial features. We introduce joint-patch and multi-label learning (JPML) to address these issues. JPML leverages group sparsity by selecting a sparse subset of facial patches while learning a multi-label classifier. In four of five comparisons on three diverse datasets, CK+, GFT, and BP4D, JPML produced the highest average F1 scores in comparison with state-of-the art.
BibTeX
@conference{Zhao-2015-5973,author = {Kaili Zhao and Wen-Sheng Chu and Fernando De la Torre Frade and Jeffrey Cohn and Honggang Zhang},
title = {Joint Patch and Multi-label Learning for Facial Action Unit Detection},
booktitle = {Proceedings of (CVPR) Computer Vision and Pattern Recognition},
year = {2015},
month = {June},
pages = {2207 - 2216},
}