期刊文献+

基于卷积与双向简单循环单元的文本分类模型 被引量:2

Text classification model based on convolution and bidirectional simple recurrent unit
在线阅读 下载PDF
导出
摘要 针对基于词粒度的长短时记忆(LSTM)网络模型存在着无法充分学习上下文语义信息的问题,提出一种基于卷积和双向简单循环单元的文本分类模型(Conv-BSA)。利用卷积和局部池化操作提取并筛选n-gram信息,使用双向简单循环单元结构提取文本深层次语义特征,通过注意力机制对深层语义特征进行加权得到最终文本表示,借助softmax函数进行分类,达到高效分辨文本类别的目的。实验结果表明,Conv-BSA模型的分类准确率高达96.09%,优于即有主流模型。简单循环单元(SRU)能够提升分类准确率,降低训练耗时。 To address the problem of being unable to fully learn long contextual semantic information in the model of word-level long short-term memory(LSTM)network,a text classification model based on convolution and bidirectional simple recurrent unit(Conv-BSA)was proposed.The n-gram information was extracted and filtered using convolution and local pooling operations.The deep semantic features of the text were extracted using the bidirectional simple recurrent unit.The deep semantic features were weighted using the attention mechanism to obtain the final text representation.The softmax function was used for classification to achieve the purpose of efficiently distinguishing text categories in the last step.Experimental results show that the classification accuracy of the Conv-BSA model is as high as 96.09%,which is better than that of the mainstream model.The simple recurrent unit(SRU)not only improves classification accuracy,but reduces training time.
作者 陈天龙 喻国平 姚磊岳 CHEN Tian-long;YU Guo-ping;YAO Lei-yue(Information Engineering School,Nanchang University,Nanchang 330031,China;Center of Collaboration and Innovation,Jiangxi University of Technology,Nanchang 330098,China)
出处 《计算机工程与设计》 北大核心 2020年第3期838-844,共7页 Computer Engineering and Design
基金 江西省科技厅科技计划基金项目(20171BBE50060) 江西省教育厅科技计划基金项目(GJJ180978) 南昌市科技局指导性科技计划基金项目(洪科字[2018]39号-73)。
关键词 卷积层 双向简单循环单元 注意力机制 文本分类 文本表示 convolutional layer bidirectional simple recurrent unit attention mechanism text classification text representation
  • 相关文献

参考文献4

二级参考文献39

  • 1CORMACK G V. Email spam filtering: a systematic review [J]. Foundations and trends in information retrieval, 2007, 1(4): 335-455.
  • 2ALMEIDA T A, YAMAKAMI A. Advances in spam filtering techniques [M]// Computational Intelligence for Privacy and Security. Berlin: Springer, 2012: 199-214.
  • 3CHOUHAN S. Behavior analysis of SVM based spam filtering using various kernel functions and data representations [J]. International journal of engineering research and technology, 2013, 2(9): 3029-3036.
  • 4PUNISKIS D, LAURUTIS R, DIRMEIKIS R. An artificial neural nets for spam E-mail recognition [J]. Electronics and electrical engineering, 2006, 69(5): 73-76.
  • 5BENGIO Y. Learning deep architectures for AI [J]. Foundations and trends in machine learning, 2009, 2(1): 1-127.
  • 6BENGIO Y, COURVILLE A, VINCENT P. Representation learning: a review and new perspectives [J]. Pattern analysis and machine intelligence, 2013, 35(8): 1798-1828.
  • 7TZORTZIS G, LIKAS A. Deep belief networks for spam filtering [C]// ICTAI 2007: Proceedings of the 19th IEEE International Conference on Tools with Artificial Intelligence. Piscataway, NJ: IEEE, 2007: 306-309.
  • 8VINCENT P, LAROCGELLE H, BENGIO Y, et al. Extracting and composing robust features with denoising autoencoders [C]// Proceedings of the 25th International Conference on Machine Learning. New York: ACM, 2008: 1096-1103.
  • 9RIFAI S, VINCENT P, MULLER X, et al. Contractive auto-encoders: explicit invariance during feature extraction [C]// Proceedings of the 28th International Conference on Machine Learning. New York: ACM, 2011: 833-840.
  • 10SRIVASTAVA N, HINTON G, KRIZHEVSKY A, et al. Dropout: a simple way to prevent neural networks from overfitting [J]. The journal of machine learning research, 2014, 15(1): 1929-1958.

共引文献82

同被引文献20

引证文献2

二级引证文献5

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部