期刊文献+

MSML-BERT模型的层级多标签文本分类方法研究 被引量:7

Study on Hierarchical Multi-Label Text Classification Method of MSML-BERT Model
下载PDF
导出
摘要 层级多标签文本分类相比普通的多标签文本分类更具有挑战性,因为文本的多个标签组织成树状的层次结构。当前方法使用相同的模型结构来预测不同层级的标签,忽略了它们之间的差异性和多样性。并且没有充分地建模层级依赖关系,造成各层级标签尤其是下层长尾标签的预测性能差,且会导致标签不一致性问题。为了解决以上问题,将多任务学习架构引入,提出了MSML-BERT模型。该模型将标签结构中每一层的标签分类网络视为一个学习任务,通过任务间知识的共享和传递,提高各层级任务的性能。基于此,设计了多尺度特征抽取模块,用于捕捉不同尺度和粒度的特征以形成不同层级需要的各种知识。进一步,设计了多层级信息传播模块,用于充分建模层级依赖,在不同层级之间传递知识,以帮助下层任务。在该模块中,设计了层次化门控机制,为了过滤不同层级任务之间的知识流动。在RCV1-V2、NYT和WOS数据集上进行了充分的实验,结果显示该模型的总体表现尤其是在下层长尾标签上的表现超过了其他主流模型,并且能维持较低的标签不一致比率。 Hierarchical multi-label text classification is more challenging than ordinary multi-label text classification,since multiple labels of the text establish a tree-like hierarchy.Current methods use the same model structure to predict labels at different layers,ignoring their differences and diversity.They don’t model the hierarchical dependencies fully,resulting in poor prediction performance of labels at all layers,especially the lower-layer long-tail labels,and may lead to label inconsistency problems.In order to address the above problems,the multi-task learning architecture is introduced,and the MSML-BERT model is proposed.The model regards the label classification network of each layer in the label hierarchy as a learning task,and enhances the performance of tasks at all layers through the sharing and transfer of knowledge between tasks.Based on this,a multi-scale feature extraction module is designed to capture multi-scale and multi-grained features to form various knowledge required at different layers.Further,a multi-layer information propagation module is designed to fully model hierarchical dependencies and transfer knowledge in different layers to support lower layer tasks.In this module,a hierarchical gating mechanism is designed to filter the knowledge flow among tasks in different layers.Extensive experiments are conducted on the RCV1-V2,NYT and WOS datasets,and the results reveal that the entire performance of this model,especially on the lower-layer long-tail labels,surpasses that of other prevailing models and maintains a low label inconsistency ratio.
作者 黄伟 刘贵全 HUANG Wei;LIU Guiquan(School of Data Science,University of Science and Technology of China,Hefei 230027,China;School of Computer Science and Technology,University of Science and Technology of China,Hefei 230027,China;Anhui Province Key Laboratory of Big Data Analysis and Application,University of Science and Technology of China,Hefei 230027,China)
出处 《计算机工程与应用》 CSCD 北大核心 2022年第15期191-201,共11页 Computer Engineering and Applications
基金 国家重点研发计划(2018YFB1801105)。
关键词 层级多标签文本分类 多任务学习架构 BERT 多尺度特征抽取模块 多层级信息传播模块 hierarchical multi-label text classification multi-task learning architecture BERT multi-scale feature extrac-tion module multi-layer information propagation module
  • 相关文献

参考文献2

二级参考文献1

共引文献13

同被引文献45

引证文献7

二级引证文献5

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部