定位
管道(软件)
宏
光流
计算机科学
人工智能
保险丝(电气)
任务(项目管理)
动作(物理)
深度学习
定位关键字
帧(网络)
计算机视觉
模式识别(心理学)
语音识别
图像(数学)
程序设计语言
物理
工程类
电气工程
经济
管理
电信
量子力学
作者
Bo Yang,Jianming Wu,Kazushi Ikeda,Gen Hattori,Masaru Sugano,Yusuke Iwasawa,Yutaka Matsuo
标识
DOI:10.1016/j.patrec.2022.12.001
摘要
This paper is an extension of our previously published ACM Multimedia 2022 paper, which was ranked 3rd in the macro-expressions (MaEs) and micro-expressions (MEs) spotting task of the FME challenge 2021. In our earlier work, a deep learning framework based on facial action units (AUs) was proposed to emphasize both local and global features to deal with the MaEs and MEs spotting tasks. In this paper, an advanced Concat-CNN model is proposed to not only utilize facial action units (AU) features, which our previous work proved were more effective in detecting MaEs, but also to fuse the optical flow features to improve the detection performance of MEs. The advanced Concat-CNN proposed in this paper not only considers the intra-features correlation of a single frame but also the inter-features correlation between frames. Further, we devise a new adaptive re-labeling method by labeling the emotional frames with distinctive scores. This method takes into account the dynamic changes in expressions to further improve the overall detection performance. Compared with our earlier work and several existing works, the newly proposed deep learning pipeline is able to achieve a better performance in terms of the overall F1-scores: 0.2623 on CAS(ME)2, 0.2839 on CAS(ME)2-cropped, and 0.3241 on SAMM-LV, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI