HiBERT: Detecting the illogical patterns with hierarchical BERT for multi-turn dialogue reasoning.

Neurocomputing(2023)

引用 1|浏览95
暂无评分
摘要
Dialogue reasoning is a new task beyond the traditional dialogue system, because it requires recognizing not only semantic relevance but also logical consistency between the candidate response and the dia-logue history. Like "all happy families are happy alike, all unhappy families are unhappy in their own way", various illogical patterns exist in the data. For example, some candidate responses use many sim-ilar words but with contradicted meanings with history; while some candidates may employ totally dif-ferent words but convey consistent meanings. Therefore, an ideal dialogue reasoning model should gather clues from both coarse-grained utterance-level and fine-grained word-level to determine the log-ical relation between candidates and the dialogue history. However, traditional models mainly rely on the widely used BERT to read all the history and candidates word by word but ignore the utterance-level sig-nals, which cannot well capture various illogical patterns in this task. To tackle this problem, we propose a novel Hierarchical BERT (HiBERT) to recognize both utterance-level and word-level illogical patterns in this paper. Specifically, BERT is firstly utilized to encode the dialogue history and each candidate response as the contextualized representation. Secondly, hierarchical reasoning architecture is conducted with this contextualized representation to obtain the word-level and the utterance-level attention distributions, respectively. In detail, we utilize the word-grained attention mechanism to obtain the word-level repre-sentation, and propose two different types of attention function, i.e, hard attention and soft attention, to obtain the utterance-grained representation. Finally, we fuse both the word-grained representation and the utterance-grained representation to calculate the logical ranking scores for the given candidate. Experimental results on two public dialogue datasets show that our model obtains higher ranking mea-sures than the widely used BERT model, validating the effectiveness of hierarchical reading of HiBERT. Further analysis on the impact of context length and attention weights shows that the HiBERT actually has the ability to recognize different illogical patterns.(c) 2022 Elsevier B.V. All rights reserved.
更多
查看译文
关键词
hierarchical hibert,dialogue,illogical patterns,multi-turn
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要