FAU, Facial Expressions, Valence and Arousal: A Multi-task Solution
Authors: Didan Deng, Zhaokang Chen, Bertram E. Shi
Abstract: In the paper, we aim to train a unified model that performs three tasks: Facial Action Units (FAU) prediction, seven basic facial expressions prediction, as well as valence and arousal prediction. The main challenge of this task is the lack of fully-annotated dataset. Most of existing datasets only contain one or two types of labels. To tackle this challenge, we propose an algorithm for the multitask model to learn from partial labels. The algorithm has two steps: first, we train a teacher model to perform all three tasks, where each instance is trained by the ground truth label of its corresponding task. Second, we refer to the outputs of the teacher model as the soft labels. We use the soft labels and the ground truths to train the student model. We find that the student model outperforms the teacher model on all the tasks, possibly due to the exposure to the full set of labels. Finally, we use ensemble modeling to boost the performance further on the three tasks.
Explore the paper tree
Click on the tree nodes to be redirected to a given paper and access their summaries and virtual assistant
Look for similar papers (in beta version)
By clicking on the button above, our algorithm will scan all papers in our database to find the closest based on the contents of the full papers and not just on metadata. Please note that it only works for papers that we have generated summaries for and you can rerun it from time to time to get a more accurate result while our database grows.