[1] World Intellectual Property Organization. World intellectual property indicators 2019[EB/OL].[2020-05-18]. https://www.wipo.int/edocs/pubdocs/en/wipo_pub_941_2019.pdf. [2] 陆晓蕾, 倪斌. 基于预训练语言模型的BERT-CNN多层级专利分类研究[EB/OL].[2020-04-06]. https://arxiv.org/ftp/arxiv/papers/1911/1911.06241.pdf. (LU X L,NI B. BERT-CNN:a hierarchical patent classifier based on a pre-trained language model[EB/OL].[2020-04-06]. https://arxiv.org/ftp/arxiv/papers/1911/1911.06241.pdf.) [3] MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality[C]//Proceedings of the 26th International Conference on Neural Information Processing Systems. Red Hook, NY:Curran Associates Inc.,2013:3111-3119. [4] MIKOLOV T,CHEN K,CORRADO G,et al. Efficient estimation of word representations in vector space[EB/OL].[2019-08-06]. http://www.surdeanu.info/mihai/teaching/ista555-spring15/readings/mikolov2013.pdf. [5] PENNINGTON J, SOCHER R, MANNING C. GloVe:global vectors for word representation[C]//Proceedings of the 2014 Conference of Empirical Methods in Natural Language Processing. Stroudsburg, PA:Association for Computational Linguistics, 2014:1532-1543. [6] PETERS M E, NEUMANN M, IYYER M, et al. Deep contextualized word representations[EB/OL].[2019-08-17]. https://arxiv.org/pdf/1802.05365.pdf. [7] RADFORD A,NARASIMHAN K,SALIMANS T,et al. Improving language understanding with unsupervised learning[EB/OL].[2019-08-06]. https://cdn.openai.com/research-covers/languageunsupervised/language_understanding_paper.pdf. [8] VASWANI A,SHAZEER N,PARMAR N,et al. Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems. Red Hook,NY:Curran Associates Inc.,2017:6000-6010. [9] DEVLIN J,CHANG M W,LEE K,et al. BERT:pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. Stroudsburg, PA:Association for Computational Linguistics,2019:4171-4186. [10] LAN Z,CHEN M,GOODMAN S,et al. ALBERT:a lite BERT for self-supervised learning of language representations[EB/OL].[2020-02-13]. https://arxiv.org/pdf/1909.11942.pdf. [11] 许甜华, 吴明礼. 一种基于TF-IDF的朴素贝叶斯算法改进[J]. 计算机技术与发展,2020,30(2):75-79.(XU T H,WU M L. An improved naive Bayes algorithm based on TF-IDF[J]. Computer Technology and Development,2020,30(2):75-79.) [12] 李程雄, 丁月华, 文贵华. SVM-KNN组合改进算法在专利文本分类中的应用[J]. 计算机工程与应用,2006,42(20):193-195,212.(LI C X,DING Y H,WEN G H. Application of SVMKNN combination improvement algorithm on patent text classification[J]. Computer Engineering and Applications,2006, 42(20):193-195,212.) [13] KIM Y. Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference of Empirical Methods in Natural Language Processing. Stroudsburg,PA:Association for Computational Linguistics,2014:1746-1751. [14] MIKOLOV T,KOMBRINK S,BURGET L,et al. Extensions of recurrent neural network language model[C]//Proceedings of the 2011 IEEE International Conference on Acoustics,Speech and Signal Processing. Piscataway:IEEE,2011:5528-5531. [15] SOCHER R,LIN C C Y,NG A Y,et al. Parsing natural scenes and natural language with recursive neural networks[C]//Proceedings of the 28th International Conference on Machine Learning. Madison,WI:Omnipress,2011:129-136. [16] HOCHREITER S,BENGIO Y,FRASCONI P. Gradient flow in recurrent nets:the difficulty of learning long-term dependencies[M]//KOLEN J F,KREMER S C. A Field Guide to Dynamical Recurrent Networks. Piscataway:IEEE,2001:237-243. [17] HOCHREITER S,SCHMIDHUBER J. Long short-term memory[J]. Neural Computation,1997,9(8):1735-1780. [18] DEY R,SALEM F M. Gate-variants of Gated Recurrent Unit (GRU) neural networks[C]//Proceedings of the IEEE 60th International Midwest Symposium on Circuits and Systems. Piscataway:IEEE,2017:1597-1600. [19] 薛金成, 姜迪, 吴建德. 基于LSTM-A深度学习的专利文本分类研究[J]. 通信技术,2019,52(12):2888-2892.(XUE J C, JIANG D,WU J D. Patent text classification based on long shortterm memory network and attention mechanism[J]. Communications Technology,2019,52(12):2888-2892.) [20] 方炯焜, 陈平华, 廖文雄. 结合GloVe和GRU的文本分类模型[J]. 计算机工程与应用,2020,56(20):98-103.(FANG J K, CHEN P H,LIAO W X. Text classification model based on GloVe and GRU[J]. Computer Engineering and Applications,2020,56(20):98-103.) [21] SUTSKEVER I,VINYALS O,LE Q V. Sequence to sequence learning with neural networks[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems. Cambridge:MIT Press,2014:3104-3112. [22] BAHDANAU D,CHO K,BENGIO Y,et al. Neural machine translation by jointly learning to align and translate[EB/OL].[2019-05-24]. https://arxiv.org/pdf/1409.0473.pdf. [23] CHO K, VAN MERRIËNBOER B, GULCEHRE C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation[C]//Proceedings of the 2014 Conference of Empirical Methods in Natural Language Processing. Stroudsburg, PA:Association for Computational Linguistics, 2014:1724-1734. [24] BA J L,KIROS J R,HINTON G E. Layer normalization[EB/OL].[2019-05-26]. https://arxiv.org/pdf/1607.06450.pdf. [25] HE K,ZHANG X,REN S,et al. Deep residual learning for image recognition[C]//Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway:IEEE, 2016:770-778. [26] YANG Z, DAI Z, YANG Y, et al. XLNet:generalized autoregressive pretraining for language understanding[EB/OL].[2020-05-26]. https://arxiv.org/pdf/1906.08237.pdf. |