论文标题

层次变压器,具有扬声器建模,以进行对话中的情感识别

A Hierarchical Transformer with Speaker Modeling for Emotion Recognition in Conversation

论文作者

Li, Jiangnan, Lin, Zheng, Fu, Peng, Si, Qingyi, Wang, Weiping

论文摘要

对话中的情感识别(ERC)比传统的文本情感识别更具挑战性。它可以被视为一项个性化和互动的情感识别任务,不仅应该考虑文本的语义信息,还应考虑说话者的影响。当前的方法通过在每两个说话者之间建立关系来模型扬声器的交互。但是,这种细粒度但复杂的建模在计算上是昂贵的,难以扩展,并且只能考虑本地环境。为了解决这个问题,我们将复杂的建模简化为二进制版本:扬声器和扬声器间的依赖项,而无需确定目标扬声器的每个独特扬声器。为了更好地实现变压器中扬声器的简化交互建模,该模型显示出了出色的解决长距离依赖性的能力,我们设计了三种类型的掩码,并分别在三个独立的变压器块中利用它们。设计的掩码分别对常规上下文建模,言论扬声器的依赖性和言论中的依赖性建模。此外,变压器块提取的不同说话者感知的信息有助于预测,因此我们利用注意机制自动加权它们。两个ERC数据集的实验表明,我们的模型有效地实现了更好的性能。

Emotion Recognition in Conversation (ERC) is a more challenging task than conventional text emotion recognition. It can be regarded as a personalized and interactive emotion recognition task, which is supposed to consider not only the semantic information of text but also the influences from speakers. The current method models speakers' interactions by building a relation between every two speakers. However, this fine-grained but complicated modeling is computationally expensive, hard to extend, and can only consider local context. To address this problem, we simplify the complicated modeling to a binary version: Intra-Speaker and Inter-Speaker dependencies, without identifying every unique speaker for the targeted speaker. To better achieve the simplified interaction modeling of speakers in Transformer, which shows excellent ability to settle long-distance dependency, we design three types of masks and respectively utilize them in three independent Transformer blocks. The designed masks respectively model the conventional context modeling, Intra-Speaker dependency, and Inter-Speaker dependency. Furthermore, different speaker-aware information extracted by Transformer blocks diversely contributes to the prediction, and therefore we utilize the attention mechanism to automatically weight them. Experiments on two ERC datasets indicate that our model is efficacious to achieve better performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源