Learning to Correct Erroneous Words for Document Grounded Conversations.

ICSCA(2023)

引用 0|浏览19
暂无评分
摘要
Document grounded conversation (DGC) aims to generate informative responses when talking about a document. It is normally formulated as a sequence-to-sequence (Seq2seq) learning problem, which directly maps source sequences, i.e., the context and background documents, to the target sequence, i.e., the response. These responses are normally used as the final output without further polishing, which may suffer from the global information loss owing to the auto-regression paradigm. To tackle this problem, some researches designed two-pass generation to improve the quality of responses. However, these approaches lack the capability of distinguishing inappropriate words in the first pass, which may maintain the erroneous words while rewrite the correct ones. In this paper, we design a scheduled error correction network (SECN) with multiple generation passes to explicitly locate and rewrite the erroneous words in previous passes. Specifically, a discriminator is employed to distinguish erroneous words which are further revised by a refiner. Moreover, we also apply curriculum learning with reasonable learning schedule to train our model from easy to hard conversations, where the complexity is measured by the number of decoding passes. We conduct comprehensive experiments on a public document grounded conversation dataset, Wizard-of-Wikipedia, and the results demonstrate significant promotions over several strong benchmarks.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要