[1] ZHANG Y,JIN R,ZHOU Z. Understanding bag-of-words model:a statistical framework[J]. International Journal of Machine Learning and Cybernetics,2010,1(1/2/3/4):43-52. [2] DAO N B,BERTET K,REVEL A. Reduction dimension of bags of visual words with FCA[C]//Proceedings of the 11th International Conference on Concept Lattices and Their Applications. Sun SITE Central Europe:CEUR-WS,2014:219-230. [3] MIKOLOV T,CHEN K,CORRADO G,et al. Efficient estimation of word representations in vector space[EB/OL].[2019-12-28]. http://arxiv.org/pdf/1301.3781.pdf. [4] 郁可人, 傅云斌, 董启文. 基于神经网络语言模型的分布式词向量研究进展[J]. 华东师范大学学报(自然科学版),2017(5):52-65,79.(YU K R,FU Y B,DONG Q W. Survey on distributed word embeddings based on neural network language models[J]. Journal of East China Normal University(Natural Science),2017(5):52-65,79.) [5] SUI H,KHOO C,CHAN S. Sentiment classification of product reviews using SVM and decision tree introduction[J]. Advances in Classification Research Online,2003,14(1):42-52. [6] HANSEN L K,SALAMON P. Neural network ensembles[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence,1990, 12(10):993-1001. [7] KIM Y. Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. Stroudsburg,PA:Associations for Computational Linguistics,2014:1746-1751. [8] LIN L,GU Z,ZHANG Z,et al. Build Chinese language model with recurrent neural network[C]//Proceedings of the 12th KIPS International Conference on Ubiquitous Information Technologies and Applications/9th International Conference on Computer Science and its Applications,LNEE 474. Cham:Springer,2017:920-925. [9] SAK H,SENIOR A W,BEAUFAYS F. Long short-term memory recurrent neural network architectures for large scale acoustic modeling[C]//Proceedings of the 15th Annual Conference of the International Speech Communication Association. Minneapolis:ISCA,2014:338-342. [10] WANG H,WEI H,GUO J,et al. Ancient Chinese sentence segmentation based on bidirectional LSTM+CRF model[J]. Journal of Advanced Computational Intelligence and Intelligent Informatics,2019,23(4):719-725. [11] ZHANG H,WANG J,ZHANG J,et al. YNU-HPCC at SemEval 2017 Task 4:using a multi-channel CNN-LSTM model for sentiment classification[C]//Proceedings of the 11th International Workshop on Semantic Evaluation. Stroudsburg,PA:Association for Computational Linguistics,2017:796-801. [12] ZHOU P,SHI W,TIAN J,et al. Attention-based bidirectional long short-term memory networks for relation classification[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, PA:Association for Computational Linguistics,2016:207-212. [13] PANKO R R. Thinking is bad:implications of human error research for spreadsheet research and practice[EB/OL].[2020-02-20]. http://arxiv.org/pdf/0801.3114.pdf. [14] VINCENT P,LAROCHELLE H,LAJOIE I,et al. Stacking denoising autoencoders:learning useful representations in a deep network with a local denoising criterion[J]. Journal of Machine Learning Research,2010,11(110):3371-3408. [15] 郭喻栋, 郭志刚, 席耀一. 基于降噪自编码器网络与词向量的信息推荐方法[J]. 计算机工程,2017,43(12):173-178. (GUO Y D,GUO Z G,XI Y Y. Information recommendation method based on denoising auto-encoder network and word vector[J]. Computer Engineering,2017,43(12):173-178.) [16] QIU X,SUN T,XU Y,et al. Pre-trained models for natural language processing:a survey[EB/OL].[2020-02-20]. https://arxiv.org/pdf/2003.08271.pdf. [17] DEVLIN J,CHANG M,LEE K,et al. BERT:pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. Stroudsburg,PA:Association for Computational Linguistics,2019:4171-4186. [18] BENGIO Y,DUCHARME R,VINCENT P,et al. A neural probabilistic language model[J]. Journal of Machine Learning Research,2003,3:1137-1155. [19] 郭宏远. 基于词向量和主题向量的文本分类算法研究[D]. 武汉:华中科技大学,2016:4-24.(GUO H Y. Text classification based on word vector and topic vector[D]. Wuhan:Huazhong University of Science and Technology,2016:4-24.) [20] VASWANI A,SHAZEER N,PARMAR N,et al. Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems. Red Hook,NY:Curran Associates Inc.,2017:6000-6010. [21] PETERS M E, NEUMANN M, LAYYER M, et al. Deep contextualized word representations[C]//Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. Stroudsburg, PA:Association for Computational Linguistics, 2018:2227-2237. [22] WU X,ZHANG T,ZANG L,et al. Mask and infill:applying masked language model to sentiment transfer[C]//Proceedings of the 28th International Joint Conference on Artificial Intelligence. Palo Alto,CA:AAAI Press,2019:5271-5277. [23] SERGIO G C, LEE M. Stacked DeBERT:all attention in incomplete data for text classification[EB/OL].[2020-05-22]. http://arxiv.org/pdf/2001.00137.pdf. |