Multi-task Hierarchical Cross-Attention Network for Multi-label Text Classification

NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II(2022)

引用 0|浏览8
暂无评分
摘要
As the quantity of scientific publications grows significantly, manual indexing of literature becomes increasingly complex, and researchers have attempted to utilize techniques in Hierarchical Multi-label Text Classification (HMTC) to classify scientific literature. Although there have been many advances, some problems still cannot be effectively solved in HMTC tasks, such as the difficulty in capturing the dependencies of hierarchical labels and the correlation between labels and text, and the lack of adaptability of models to specialized text. In this paper, we propose a novel framework called Multi-task Hierarchical Cross-Attention Network (MHCAN) for multi-label text classification. Specifically, we introduce a cross-attention mechanism to fully incorporate text representation and hierarchical labels with a directed acyclic graph (DAG) structure, and design an iterative hierarchical-attention module to capture the dependencies between layers. Afterwards, our framework weighting jointly optimizes each level of loss. To improve the adaptability of the model to domain data, we also continue to pre-train SciBERT on unlabeled data and introduce adversarial training. Our framework ranks 2nd in NLPCC 2022 Shared Task 5 Track 1 (Multi-label Classification Model for English Scientific Literature). The experimental results show the effectiveness of the modules applied in this framework.
更多
查看译文
关键词
Hierarchical multi-label text classification, Multi-task learning, Attention mechanism
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要