Journal of Computer Applications ›› 2022, Vol. 42 ›› Issue (4): 1116-1124.DOI: 10.11772/j.issn.1001-9081.2021071257
• The 36 CCF National Conference of Computer Applications (CCF NCCA 2020) • Previous Articles
Haifeng ZHANG1, Cheng ZENG1,2,3(), Lie PAN1, Rusong HAO1, Chaodong WEN1, Peng HE1,2,3
Received:
2021-07-16
Revised:
2021-11-11
Accepted:
2021-11-17
Online:
2022-04-15
Published:
2022-04-10
Contact:
Cheng ZENG
About author:
ZHANG Haifeng, born in 1990, M. S. candidate. His research interests include natural language processing, text classification.Supported by:
张海丰1, 曾诚1,2,3(), 潘列1, 郝儒松1, 温超东1, 何鹏1,2,3
通讯作者:
曾诚
作者简介:
张海丰(1990—),男,湖北黄冈人,硕士研究生,主要研究方向:自然语言处理、文本分类基金资助:
CLC Number:
Haifeng ZHANG, Cheng ZENG, Lie PAN, Rusong HAO, Chaodong WEN, Peng HE. News topic text classification method based on BERT and feature projection network[J]. Journal of Computer Applications, 2022, 42(4): 1116-1124.
张海丰, 曾诚, 潘列, 郝儒松, 温超东, 何鹏. 结合BERT和特征投影网络的新闻主题文本分类方法[J]. 《计算机应用》唯一官方网站, 2022, 42(4): 1116-1124.
Add to citation manager EndNote|Ris|BibTeX
URL: http://www.joca.cn/EN/10.11772/j.issn.1001-9081.2021071257
名称 | 值 |
---|---|
CPU | Intel Xeon Gold 5218 |
GPU | NVIDIA GeForce RTX5000-16G |
开发语言 | Python-3. 6 |
深度学习框架 | Pytorch-1.2.0 |
开发工具 | Pycharm-2020.1.3 |
Tab. 1 Experimental environment
名称 | 值 |
---|---|
CPU | Intel Xeon Gold 5218 |
GPU | NVIDIA GeForce RTX5000-16G |
开发语言 | Python-3. 6 |
深度学习框架 | Pytorch-1.2.0 |
开发工具 | Pycharm-2020.1.3 |
数据集 | 类别 | 平均 长度 | 样本数 | |||
---|---|---|---|---|---|---|
总数 | 训练集 | 验证集 | 测试集 | |||
今日头条 | 15 | 22 | 382 688 | 267 878 | 57 409 | 57 401 |
搜狐新闻 | 12 | 17 | 34 218 | 22 699 | 5 755 | 5 764 |
THUCNews-L | 10 | 19 | 200 000 | 180 000 | 10 000 | 10 000 |
THUCNews-S | 6 | 18 | 60 000 | 48 000 | 6 000 | 6 000 |
Tab. 2 Dataset details
数据集 | 类别 | 平均 长度 | 样本数 | |||
---|---|---|---|---|---|---|
总数 | 训练集 | 验证集 | 测试集 | |||
今日头条 | 15 | 22 | 382 688 | 267 878 | 57 409 | 57 401 |
搜狐新闻 | 12 | 17 | 34 218 | 22 699 | 5 755 | 5 764 |
THUCNews-L | 10 | 19 | 200 000 | 180 000 | 10 000 | 10 000 |
THUCNews-S | 6 | 18 | 60 000 | 48 000 | 6 000 | 6 000 |
名称 | 值 | 名称 | 值 |
---|---|---|---|
hidden_size | 768 | vocab_size | 21 128 |
num_attention_heads | 12 | hidden_act | Gelu |
num_hidden_layers | 12 |
Tab. 3 Major parameters of BERT model
名称 | 值 | 名称 | 值 |
---|---|---|---|
hidden_size | 768 | vocab_size | 21 128 |
num_attention_heads | 12 | hidden_act | Gelu |
num_hidden_layers | 12 |
名称 | 值 | 名称 | 值 |
---|---|---|---|
optimizer | BertAdam | batchsize | 128 |
warmup | 0.1 | λ | |
learningrate | 5E-5 | Dropout | 0.5 |
pad_size | 32 |
Tab. 4 Hyperparameters of BERT-FPnet model
名称 | 值 | 名称 | 值 |
---|---|---|---|
optimizer | BertAdam | batchsize | 128 |
warmup | 0.1 | λ | |
learningrate | 5E-5 | Dropout | 0.5 |
pad_size | 32 |
特征投影层 | 搜狐新闻 | ||
---|---|---|---|
Acc | M_F1 | ||
单层投影 | 3 | 0.824 3 | 0.823 2 |
6 | 0.845 9 | 0.846 6 | |
9 | 0.843 1 | 0.844 2 | |
12 | 0.861 7 | 0.862 7 | |
双层投影 | 3-MLP | 0.838 6 | 0.838 4 |
6-MLP | 0.852 5 | 0.853 1 | |
9-MLP | 0.840 0 | 0.842 9 | |
12-MLP | 0.858 9 | 0.860 4 | |
所有层投影 | ALL | 0.813 7 | 0.812 7 |
方 | BERT-FPnet-1 | 0.852 5 | 0.852 7 |
Tab. 5 Experimental results of BERT-FPnet-2 hidden layer feature projection on Sohu News dataset
特征投影层 | 搜狐新闻 | ||
---|---|---|---|
Acc | M_F1 | ||
单层投影 | 3 | 0.824 3 | 0.823 2 |
6 | 0.845 9 | 0.846 6 | |
9 | 0.843 1 | 0.844 2 | |
12 | 0.861 7 | 0.862 7 | |
双层投影 | 3-MLP | 0.838 6 | 0.838 4 |
6-MLP | 0.852 5 | 0.853 1 | |
9-MLP | 0.840 0 | 0.842 9 | |
12-MLP | 0.858 9 | 0.860 4 | |
所有层投影 | ALL | 0.813 7 | 0.812 7 |
方 | BERT-FPnet-1 | 0.852 5 | 0.852 7 |
特征投影层 | THUCNews-S | |
---|---|---|
Acc | M_F1 | |
12 | 0.936 2 | 0.936 0 |
BERT-FPnet-1 | 0.937 3 | 0.937 2 |
Tab. 6 Comparison of BERT-FPnet feature projection results on THUCNews-S dataset
特征投影层 | THUCNews-S | |
---|---|---|
Acc | M_F1 | |
12 | 0.936 2 | 0.936 0 |
BERT-FPnet-1 | 0.937 3 | 0.937 2 |
词嵌入 | 模型 | 今日头条 | 搜狐新闻 | THUCNews-L | THUCNews-S | ||||
---|---|---|---|---|---|---|---|---|---|
Acc | M_F1 | Acc | M_F1 | Acc | M_F1 | Acc | M_F1 | ||
Word2Vec | TextCNN | 0.832 1 | 0.767 8 | 0.832 0 | 0.833 3 | 0.910 5 | 0.910 7 | 0.900 8 | 0.900 5 |
FastText | 0.839 3 | 0.773 3 | 0.823 6 | 0.823 6 | 0.920 8 | 0.920 9 | 0.898 2 | 0.898 3 | |
Transformer | 0.793 9 | 0.733 7 | 0.781 6 | 0.781 4 | 0.897 3 | 0.897 1 | 0.884 5 | 0.884 3 | |
DPCNN | 0.816 8 | 0.754 4 | 0.770 5 | 0.769 8 | 0.907 6 | 0.907 6 | 0.898 3 | 0.898 3 | |
ALBERT | ALBERT-FC | 0.846 0 | 0.782 9 | 0.837 5 | 0.838 4 | 0.926 0 | 0.926 3 | 0.910 5 | 0.910 2 |
BERT | BERT-FC | 0.855 9 | 0.791 2 | 0.842 2 | 0.841 6 | 0.932 5 | 0.932 4 | 0.922 7 | 0.922 8 |
BERT-CNN | 0.862 0 | 0.796 5 | 0.847 3 | 0.848 6 | 0.942 1 | 0.942 1 | 0.935 3 | 0.935 0 | |
BERT-BIGRU | 0.862 4 | 0.798 1 | 0.845 9 | 0.847 2 | 0.935 2 | 0.935 2 | 0.926 2 | 0.926 2 | |
BERT-FPnet-1 | 0.869 6 | 0.803 1 | 0.852 5 | 0.852 7 | 0.944 0 | 0.943 8 | 0.937 3 | 0.937 2 | |
BERT-FPnet-2 | 0.868 0 | 0.801 1 | 0.861 7 | 0.862 7 | 0.941 0 | 0.942 3 | 0.936 2 | 0.936 0 |
Tab. 7 Experimental results of different models on different datasets
词嵌入 | 模型 | 今日头条 | 搜狐新闻 | THUCNews-L | THUCNews-S | ||||
---|---|---|---|---|---|---|---|---|---|
Acc | M_F1 | Acc | M_F1 | Acc | M_F1 | Acc | M_F1 | ||
Word2Vec | TextCNN | 0.832 1 | 0.767 8 | 0.832 0 | 0.833 3 | 0.910 5 | 0.910 7 | 0.900 8 | 0.900 5 |
FastText | 0.839 3 | 0.773 3 | 0.823 6 | 0.823 6 | 0.920 8 | 0.920 9 | 0.898 2 | 0.898 3 | |
Transformer | 0.793 9 | 0.733 7 | 0.781 6 | 0.781 4 | 0.897 3 | 0.897 1 | 0.884 5 | 0.884 3 | |
DPCNN | 0.816 8 | 0.754 4 | 0.770 5 | 0.769 8 | 0.907 6 | 0.907 6 | 0.898 3 | 0.898 3 | |
ALBERT | ALBERT-FC | 0.846 0 | 0.782 9 | 0.837 5 | 0.838 4 | 0.926 0 | 0.926 3 | 0.910 5 | 0.910 2 |
BERT | BERT-FC | 0.855 9 | 0.791 2 | 0.842 2 | 0.841 6 | 0.932 5 | 0.932 4 | 0.922 7 | 0.922 8 |
BERT-CNN | 0.862 0 | 0.796 5 | 0.847 3 | 0.848 6 | 0.942 1 | 0.942 1 | 0.935 3 | 0.935 0 | |
BERT-BIGRU | 0.862 4 | 0.798 1 | 0.845 9 | 0.847 2 | 0.935 2 | 0.935 2 | 0.926 2 | 0.926 2 | |
BERT-FPnet-1 | 0.869 6 | 0.803 1 | 0.852 5 | 0.852 7 | 0.944 0 | 0.943 8 | 0.937 3 | 0.937 2 | |
BERT-FPnet-2 | 0.868 0 | 0.801 1 | 0.861 7 | 0.862 7 | 0.941 0 | 0.942 3 | 0.936 2 | 0.936 0 |
模型 | pad_size | THUCNews-S | |
---|---|---|---|
Acc | F1值 | ||
BERT-FPnet-1 | 18 | 0.928 7 | 0.927 8 |
24 | 0.932 0 | 0.931 9 | |
32 | 0.937 3 | 0.937 2 | |
40 | 0.930 5 | 0.930 5 | |
BERT-FPnet-2 | 18 | 0.931 6 | 0.930 7 |
24 | 0.935 7 | 0.934 6 | |
32 | 0.936 2 | 0.936 0 | |
40 | 0.936 2 | 0.936 1 |
Tab. 8 Perfomance comparison of proposed models under different pad_size on THUCNews-S dataset
模型 | pad_size | THUCNews-S | |
---|---|---|---|
Acc | F1值 | ||
BERT-FPnet-1 | 18 | 0.928 7 | 0.927 8 |
24 | 0.932 0 | 0.931 9 | |
32 | 0.937 3 | 0.937 2 | |
40 | 0.930 5 | 0.930 5 | |
BERT-FPnet-2 | 18 | 0.931 6 | 0.930 7 |
24 | 0.935 7 | 0.934 6 | |
32 | 0.936 2 | 0.936 0 | |
40 | 0.936 2 | 0.936 1 |
模型 | λ | Acc | F1值 |
---|---|---|---|
BERT- FPnet-1 | 1 | 0.934 3 | 0.934 6 |
[0.25,0.5,0.75,1] | 0.937 3 | 0.937 4 | |
[0.05, 0.1, 0.2, 0.4, 0.8, 1.0] | 0.937 3 | 0.937 2 | |
BERT- FPnet-2 | 1 | 0.936 2 | 0.936 0 |
[0.25,0.5,0.75,1] | 0.933 3 | 0.933 5 | |
[0.05, 0.1, 0.2, 0.4, 0.8, 1.0] | 0.936 2 | 0.936 0 |
Tab. 9 Performance comparison of proposed models under different λ on THUCNews-S dataset
模型 | λ | Acc | F1值 |
---|---|---|---|
BERT- FPnet-1 | 1 | 0.934 3 | 0.934 6 |
[0.25,0.5,0.75,1] | 0.937 3 | 0.937 4 | |
[0.05, 0.1, 0.2, 0.4, 0.8, 1.0] | 0.937 3 | 0.937 2 | |
BERT- FPnet-2 | 1 | 0.936 2 | 0.936 0 |
[0.25,0.5,0.75,1] | 0.933 3 | 0.933 5 | |
[0.05, 0.1, 0.2, 0.4, 0.8, 1.0] | 0.936 2 | 0.936 0 |
模型 | 双网络策略 | Acc | F1值 |
---|---|---|---|
BERT-FPnet-1 | 同步 | 0.937 3 | 0.937 2 |
异步 | 0.932 3 | 0.932 3 | |
BERT-FPnet-2 | 同步 | 0.936 2 | 0.936 0 |
异步 | 0.935 3 | 0.935 4 |
Tab. 10 Performance comparison of proposed models under different optimization strategies on THUCNews-S dataset
模型 | 双网络策略 | Acc | F1值 |
---|---|---|---|
BERT-FPnet-1 | 同步 | 0.937 3 | 0.937 2 |
异步 | 0.932 3 | 0.932 3 | |
BERT-FPnet-2 | 同步 | 0.936 2 | 0.936 0 |
异步 | 0.935 3 | 0.935 4 |
1 | 许甜华,吴明礼. 一种基于TF-IDF的朴素贝叶斯算法改进[J]. 计算机技术与发展, 2020, 30(2):75-79. 10.3969/j.issn.1673-629X.2020.02.016 |
XU T H, WU M L. An improved naive Bayes algorithm based on TF-IDF[J]. Computer Technology and Development, 2020, 30(2): 75-79. 10.3969/j.issn.1673-629X.2020.02.016 | |
2 | MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality[C]// Proceedings of the 26th International Conference on Neural Information Processing Systems. Red Hook, NY: Curran Associates Inc., 2013: 3111-3119. |
3 | MIKOLOV T, CHEN K, CORRADO G, et al. Efficient estimation of word representations in vector space[EB/OL]. (2013-09-07) [2020-08-06].. 10.3126/jiee.v3i1.34327 |
4 | PENNINGTON J, SOCHER R, MANNING C D. GloVe: global vectors for word representation[C]// Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA: Association for Computational Linguistics, 2014: 1532-1543. 10.3115/v1/d14-1162 |
5 | 李舟军,范宇,吴贤杰. 面向自然语言处理的预训练技术研究综述[J]. 计算机科学, 2020, 47(3):162-173. 10.11896/jsjkx.191000167 |
LI Z J, FAN Y, WU X J. Survey of natural language processing pre-training techniques[J]. Computer Science, 2020, 47(3):162-173. 10.11896/jsjkx.191000167 | |
6 | DEVLIN J, CHANG M W, LEE K, et al. BERT: pre-training of deep bidirectional transformers for language understanding[C]// Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Stroudsburg, PA: Association for Computational Linguistics, 2019: 4171-4186. 10.18653/v1/n19-1423 |
7 | KIM Y. Convolutional neural networks for sentence classification[C]// Proceedings of the 2014 Conference of Empirical Methods in Natural Language Processing. Stroudsburg, PA: Association for Computational Linguistics, 2014: 1746-1751. 10.3115/v1/d14-1181 |
8 | ZHANG X, ZHAO J B, LeCUN Y. Character-level convolutional networks for text classification[C]// Proceedings of the 28th International Conference on Neural Information Processing Systems. Cambridge: MIT Press, 2015: 649-657. 10.1109/icip.2015.7351229 |
9 | MIKOLOV T, KOMBRINK S, BURGET L, et al. Extensions of recurrent neural network language model[C]// Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing. Piscataway: IEEE, 2011: 5528-5531. 10.1109/icassp.2011.5947611 |
10 | SOCHER R, LIN C C Y, NG A Y, et al. Parsing natural scenes and natural language with recursive neural networks[C]// Proceedings of the 28th International Conference on Machine Learning. Madison, WI: Omnipress, 2011: 129-136. |
11 | 何力,郑灶贤,项凤涛,等. 基于深度学习的文本分类技术研究进展[J]. 计算机工程, 2021, 47(2):1-11. |
HE L, ZHENG Z X, XIANG F T, et al. Research progress of text classification technology based on deep learning[J]. Computer Engineering, 2021, 47(2):1-11. | |
12 | LAI S W, XU L H, LIU K, et al. Recurrent convolutional neural networks for text classification[C]// Proceedings of the 29th AAAI Conference on Artificial Intelligence. Palo Alto, CA: AAAI Press, 2015: 2267-2273. 10.1609/aaai.v33i01.33017370 |
13 | XIAO Y J, CHO K. Efficient character-level document classification by combining convolution and recurrent layers[EB/OL]. (2016-02-01) [2021-05-25].. |
14 | VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[C]// Proceedings of the 31st International Conference of Neural Information Processing Systems. Red Hook, NY: Curran Associates Inc., 2017: 6000-6010. 10.1016/s0262-4079(17)32358-8 |
15 | ZHOU P, SHI W, TIAN J, et al. Attention-based bidirectional long short-term memory networks for relation classification[C]// Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Stroudsburg, PA: Association for Computational Linguistics, 2016:207-212. 10.18653/v1/p16-2034 |
16 | 杨春霞,李锐,秦家鹏. 一种粒度融合的新闻文本主题分类模型[J]. 小型微型计算机系统, 2020, 41(11):2256-2259. 10.3969/j.issn.1000-1220.2020.11.003 |
YANG C X, LI R, QIN J P. Granular fusion news text topic classification model[J]. Journal of Chinese Computer Systems, 2020, 41(11):2256-2259. 10.3969/j.issn.1000-1220.2020.11.003 | |
17 | 付静,龚永罡,廉小亲,等. 基于BERT-LDA的新闻短文本分类方法[J]. 信息技术与信息化, 2021(2):127-129. 10.3969/j.issn.1672-9528.2021.02.044 |
FU J, GONG Y G, LIAN X Q, et al. News short text classification method based on BERT-LDA[J]. Information Technology and Informatization, 2021(2):127-129. 10.3969/j.issn.1672-9528.2021.02.044 | |
18 | LAN Z Z, CHEN M D, GOODMAN S, et al. ALBERT: a lite BERT for self-supervised learning of language representations[EB/OL]. (2020-02-09) [2021-05-25].. |
19 | 温超东,曾诚,任俊伟,等. 结合ALBERT和双向门控循环单元的专利文本分类[J]. 计算机应用, 2021, 41(2):407-412. |
WEN C D, ZENG C, REN J W, et al. Patent text classification combining ALBERT and bidirectional gated recurrent unit[J]. Journal of Computer Applications, 2021, 41(2):407-412. | |
20 | CHEN J A, YANG Z C, YANG D Y. MixText: linguistically-informed interpolation of hidden space for semi-supervised text classification[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, PA: Association for Computational Linguistics, 2020:2147-2157. 10.18653/v1/2020.acl-main.194 |
21 | MENG Y, ZHANG Y Y, HUANG J X, et al. Text classification using label names only: a language model self-training approach[C]// Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA: Association for.Computational Linguistics, 2020:9006-9017. 10.18653/v1/2020.emnlp-main.724 |
22 | QIN Q, HU W P, LIU B. Feature projection for improved text classification[C]// Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, PA: Association for Computational Linguistics, 2020: 8161-8171. 10.18653/v1/2020.acl-main.726 |
23 | GANIN Y, LEMPITSKY V. Unsupervised domain adaptation by backpropagation[C]// Proceedings of the 32nd International Conference on Machine Learning. New York: JMLR.org, 2015: 1180-1189. |
24 | GANIN Y, USTINOVA E, AJAKAN H, et al. Domain-adversarial training of neural networks[J]. Journal of Machine Learning Research, 2016, 17: 1-35. 10.1007/978-3-319-58347-1_10 |
25 | ZHANG K, ZHANG H F, LIU Q, et al. Interactive attention transfer network for cross-domain sentiment classification[C]// Proceedings of the 33rd AAAI Conference on Artificial Intelligence. Palo Alto, CA: AAAI Press, 2019: 5773-5780. 10.1609/aaai.v33i01.33015773 |
26 | JAWAHAR G, SAGOT B, SEDDAH D. What does BERT learn about the structure of language?[C]// Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Stroudsburg, PA: Association for Computational Linguistics, 2019:3651-3657. 10.18653/v1/p19-1356 |
27 | JOULIN A, GRAVE E, BOJANOWSKI P, et al. Bag of tricks for efficient text classification[C]// Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2, Short Papers). Stroudsburg, PA: Association for Computational Linguistics, 2017: 427-431. 10.18653/v1/e17-2068 |
28 | JOHNSON R, ZHANG T. Deep pyramid convolutional neural networks for text categorization[C]// Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Stroudsburg, PA: Association for Computational Linguistics, 2017: 562-570. 10.18653/v1/p17-1052 |
29 | HE K M, ZHANG X Y, REN S Q, et al. Deep residual learning for image recognition[C]// Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2016: 770-778. 10.1109/cvpr.2016.90 |
30 | 史振杰,董兆伟,庞超逸,等. 基于BERT-CNN的电商评论情感分析[J]. 智能计算机与应用, 2020, 10(2):7-11. 10.3969/j.issn.2095-2163.2020.02.002 |
SHI Z J, DONG Z W, PANG C Y, et al. Sentiment analysis of e-commerce reviews based on BERT-CNN[J]. Intelligent Computer and Applications, 2020, 10(2):7-11. 10.3969/j.issn.2095-2163.2020.02.002 | |
31 | YU Q, WANG Z Y, JIANG K W. Research on text classification based on BERT-BiGRU model[J]. Journal of Physics: Conference Series, 2021, 1746: No.012019. 10.1088/1742-6596/1746/1/012019 |
[1] | Wangjing TANG, Bin XU, Meihan TONG, Meihuan HAN, Liming WANG, Qi ZHONG. Popular science text classification model enhanced by knowledge graph [J]. Journal of Computer Applications, 2022, 42(4): 1072-1078. |
[2] | Shoulong JIAO, Youxiang DUAN, Qifeng SUN, Zihao ZHUANG, Chenhao SUN. Knowledge representation learning method incorporating entity description information and neighbor node features [J]. Journal of Computer Applications, 2022, 42(4): 1050-1056. |
[3] | Zengzhen DU, Dongxin TANG, Dan XIE. Method of generating rhetorical questions based on deep neural network in intelligent consultation [J]. Journal of Computer Applications, 2022, 42(3): 867-873. |
[4] | Xueqiang LYU, Chen PENG, Le ZHANG, Zhi’an DONG, Xindong YOU. Text multi-label classification method incorporating BERT and label semantic attention [J]. Journal of Computer Applications, 2022, 42(1): 57-63. |
[5] | Qiming RUAN, Yi GUO, Nan ZHENG, Yexiang WANG. Customs declaration good classification algorithm based on hierarchical multi-task BERT [J]. Journal of Computer Applications, 2022, 42(1): 71-77. |
[6] | ZHANG Yang, JIANG Minghu. Authorship identification of text based on attention mechanism [J]. Journal of Computer Applications, 2021, 41(7): 1897-1901. |
[7] | WEN Chaodong, ZENG Cheng, REN Junwei, ZHANG Yan. Patent text classification based on ALBERT and bidirectional gated recurrent unit [J]. Journal of Computer Applications, 2021, 41(2): 407-412. |
[8] | Yang ZHANG, Xiaoning WANG. Text feature selection method based on Word2Vec word embedding and genetic algorithm for biomarker selection in high-dimensional omics [J]. Journal of Computer Applications, 2021, 41(11): 3151-3155. |
[9] | Zhichao LI, Tohti TURDI, Hamdulla ASKAR. Answer selection model based on dynamic attention and multi-perspective matching [J]. Journal of Computer Applications, 2021, 41(11): 3156-3163. |
[10] | TAN Jinyuan, DIAO Yufeng, QI Ruihua, LIN Hongfei. Automatic summary generation of Chinese news text based on BERT-PGN model [J]. Journal of Computer Applications, 2021, 41(1): 127-132. |
[11] | LIAO Shenglan, YIN Shi, CHEN Xiaoping, ZHANG Bo, OUYANG Yu, ZHANG Heng. Intent recognition dataset for dialogue systems in power business [J]. Journal of Computer Applications, 2020, 40(9): 2549-2554. |
[12] | YIN Chunyong, HE Miao. Text classification based on improved capsule network [J]. Journal of Computer Applications, 2020, 40(9): 2525-2530. |
[13] | CHEN Jiawei, HAN Fang, WANG Zhijie. Aspect-based sentiment analysis with self-attention gated graph convolutional network [J]. Journal of Computer Applications, 2020, 40(8): 2202-2206. |
[14] | WANG Minrui, GAO Shu, YUAN Ziyong, YUAN Lei. Sequence generation model with dynamic routing for multi-label text classification [J]. Journal of Computer Applications, 2020, 40(7): 1884-1890. |
[15] | YANG Guofeng, YANG Yong. Question classification of common crop disease question answering system based on BERT [J]. Journal of Computer Applications, 2020, 40(6): 1580-1586. |
Viewed | ||||||
Full text |
|
|||||
Abstract |
|
|||||