论文标题

元转移学习以识别情绪

Meta Transfer Learning for Emotion Recognition

论文作者

Nguyen, Dung, Sridharan, Sridha, Nguyen, Duc Thanh, Denman, Simon, Dean, David, Fookes, Clinton

论文摘要

深度学习在自动情绪识别方面已被广泛采用,并在该领域取得了重大进展。但是,由于注释的情绪数据集不足,预训练的模型的概括能力受到限制,因此导致新型测试集的性能不佳。为了减轻这一挑战,已经应用了对预训练模型的转移学习进行微调。但是,微调的知识可能会覆盖和/或丢弃从预先训练的模型中学到的重要知识。在本文中,我们通过提出一种基于路径网的传输学习方法来解决这个问题,该方法能够将从一个视觉/音频情感域中学到的情感知识转移到另一个视觉/音频情感域,并将从多个音频情感界所学的情感知识转移到彼此之间,以提高整体情感识别的准确性。为了显示我们提出的系统的鲁棒性,在三个情绪数据集上进行了各种面部表达识别和语音情感识别任务的实验集:Savee,EMODB和Enterface已进行。实验结果表明,我们所提出的系统能够改善情绪识别的性能,从而使其性能优于最近提出的基于基于基于的转移学习方法的微调/预训练/预训练的模型。

Deep learning has been widely adopted in automatic emotion recognition and has lead to significant progress in the field. However, due to insufficient annotated emotion datasets, pre-trained models are limited in their generalization capability and thus lead to poor performance on novel test sets. To mitigate this challenge, transfer learning performing fine-tuning on pre-trained models has been applied. However, the fine-tuned knowledge may overwrite and/or discard important knowledge learned from pre-trained models. In this paper, we address this issue by proposing a PathNet-based transfer learning method that is able to transfer emotional knowledge learned from one visual/audio emotion domain to another visual/audio emotion domain, and transfer the emotional knowledge learned from multiple audio emotion domains into one another to improve overall emotion recognition accuracy. To show the robustness of our proposed system, various sets of experiments for facial expression recognition and speech emotion recognition task on three emotion datasets: SAVEE, EMODB, and eNTERFACE have been carried out. The experimental results indicate that our proposed system is capable of improving the performance of emotion recognition, making its performance substantially superior to the recent proposed fine-tuning/pre-trained models based transfer learning methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源