Learning facial expression-aware global-to-local representation for robust action unit detection.

Autor: An, Rudong, Jin, Aobo, Chen, Wei, Zhang, Wei, Zeng, Hao, Deng, Zhigang, Ding, Yu
Předmět:
Zdroj: Applied Intelligence; Jan2024, Vol. 54 Issue 2, p1405-1425, 21p
Abstrakt: The task of detecting facial action units (AU) often utilizes discrete expression categories, such as Angry, Disgust, and Happy, as auxiliary information to enhance performance. However, these categories are unable to capture the subtle transformations of AUs. Additionally, existing works suffer from overfitting due to the limited availability of AU datasets. This paper proposes a novel fine-grained global expression representation encoder to capture continuous and subtle global facial expressions and improve AU detection. The facial expression representation effectively reduces overfitting by isolating facial expressions from other factors such as identity, background, head pose, and illumination. To further address overfitting, a local AU features module transforms the global expression representation into local facial features for each AU. Finally, the local AU features are fed into an AU classifier to determine the occurrence of each AU. Our proposed method outperforms previous works and achieves state-of-the-art performances on both in-the-lab and in-the-wild datasets. This is in contrast to most existing works that only focus on in-the-lab datasets. Our method specifically addresses the issue of overfitting from limited data, which contributes to its superior performance. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index