SSER: Semi-Supervised Emotion Recognition based on Triplet Loss and pseudo label Article

Pan, L, Shao, W, Xiong, S et al. (2024). SSER: Semi-Supervised Emotion Recognition based on Triplet Loss and pseudo label . KNOWLEDGE-BASED SYSTEMS, 292 10.1016/j.knosys.2024.111595

cited authors

  • Pan, L; Shao, W; Xiong, S; Lei, Q; Huang, S; Beckman, E; Hu, Q

authors

abstract

  • Recently, emotion recognition from facial expressions has achieved unprecedented accuracy with the development of deep learning. Despite this progress, most existing emotion recognition methods are supervised and thus require extensive annotation. This issue is particularly pronounced in continuous domain datasets where annotation costs are very high. Furthermore, discrete domain datasets containing specific poses are too uniform to reflect complex and actual emotions. Existing methods that employ classification loss pay little attention to image similarity, making it difficult to distinguish similar emotions. To improve the learning ability for image similarity and reduce the annotation cost of continuous domain datasets, this research proposes a Semi-Supervised Emotion Recognition (SSER) method, which incorporates Activation-matrix Triplet loss (AMT loss) and pseudo label with Complementary Information (CI label). Specifically, the AMT loss is constructed by encoding multiple activation channels of an image as a matrix, which are utilized to capture the image similarity. The CI label firstly adopts the coupling effect of the complementary information from images and the multi-stage model for SSL to obtain high-confidence pseudo-labels. Then, entropy minimization and consistency regularization are used to improve the accuracy of pseudo labels. The SSER is evaluated on continuous domain datasets (AFEW-VA and AFF-Wild) and discrete domain datasets (FER2013 and CK+). The experimental results demonstrate that the SSER combined with AMT loss and CI label makes improvement for emotion recognition on continuous domain datasets, meanwhile the SSER is also desirable and effective for emotion recognition on discrete domain datasets.

publication date

  • May 23, 2024

published in

Digital Object Identifier (DOI)

volume

  • 292