[1] CHEN K T,CHANG C J,WU C C,et al. Quadrant of euphoria:a crowdsourcing platform for QoE assessment[J]. IEEE Network,2010,24(2):28-35. [2] KOHLER T. Crowdsourcing-based business models[J]. California Management Review,2015,57(4):63-84. [3] SCHUTTE N S, MALOUFF J M, THORSTEINSSON E B. Increasing emotional intelligence through training:current status and future directions[J]. International Journal of Emotional Education,2013,5(1):56-72. [4] 韩文静, 李海峰, 阮华斌, 等. 语音情感识别研究进展综述[J]. 软件学报,2014,25(1):37-50.(HAN W J,LI H F,RUAN H B,et al. Review on speech emotion recognition[J]. Journal of Software, 2014,25(1):37-50.) [5] LOCATELLO F,BAUER S,LUCIC M,et al. Challenging common assumptions in the unsupervised learning of disentangled representations[J]. Statistics,2019,2(6):238-249. [6] WILLIS C G,LAW E,WILLIAMS A C,et al. CrowdCurio:an online crowdsourcing platform to facilitate climate change studies using herbarium specimens[J]. New Phytologist,2017,215(1):479-488. [7] 李宏言, 范利春, 高鹏, 等. 大数据语音语料库的社会标注技术[J]. 清华大学学报(自然科学版),2013,53(6):908-912.(LI H Y,FAN L C,GAO P,et al. Social annotation for large speech corpora[J]. Journal of Tsinghua University (Natural Science Edition),2013,53(6):908-912.) [8] FU K,LI J,JIN J,et al. Image-text surgery:efficient concept learning in image captioning by generating pseudopairs[J]. IEEE Transactions on Neural Networks and Learning Systems,2018,29(12):5910-5921. [9] GAO L,FAN K,SONG J,et al. Deliberate attention networks for image captioning[C]//Proceedings of the 33rd AAAI Conference on Artificial Intelligence. Palo Alto,CA:AAAI,2019:8320-8327. [10] CHEN F,JI R,SUN X,et al. GroupCap:group-based image captioning with structured relevance and diversity constraints[C]//Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway:IEEE,2018:1345-1353. [11] 徐世鹏, 杨鸿武, 王海燕. 面向藏语语音合成的语音基元自动标注方法[J]. 计算机工程与应用,2015,51(6):199-203.(XU S P,YANG H W,WANG H Y. Speech unit segmentation for Tibetan speech synthesis[J]. Computer Engineering and Applications,2015,51(6):199-203.) [12] 唐素勤, 孙亚茹, 李志欣, 等. 基于强化学习的壮语词性标注[J]. 计算机工程,2020,46(4):309-315.(TANG S Q,SUN Y R,LI Z X,et al. Part of speech tagging of Zhuang language based on reinforcement learning[J]. Computer Engineering,2020,46(4):309-315.) [13] 傅睿博, 陶建华, 李雅, 等. 基于静音时长和文本特征融合的韵律边界自动标注[J]. 清华大学学报(自然科学版),2018,58(1):61-66,74.(FU R B,TAO J H,LI Y,et al. Automatic prosodic boundary labeling based on fusing the silence duration with the lexical features[J]. Journal of Tsinghua University (Science and Technology),2018,58(1):61-66,74.) [14] COWIE R, DOUGLAS-COWIE E, SAVVIDOU S, et al. ‘FEELTRACE’:an instrument for recording perceived emotion in real time[C]//Proceedings of the 2000 ISCA Tutorial and Research Workshop on Speech and Emotion. Belfast:International Speech Communication Association,2000:19-24. [15] EYBEN F,WOLLMER M,SCHULLER B. openSMILE-the Munich versatile and fast open-source audio feature extractor[C]//Proceedings of the 18th ACM International Conference on Multimedia. New York:ACM,2010:1459-1462. [16] MCKEOWN G,VALSTAR M F,COWIE R,et al. The SEMAINE corpus of emotionally coloured character interactions[C]//Proceedings of the 2010 IEEE International Conference on Multimedia and Expo. Piscataway:IEEE,2010:1079-1084. [17] 陈盼弟, 黄华, 何凌. 基于自相关和倒谱法的基音检测改进算法[J]. 计算机应用与软件,2015,32(1):163-166.(CHENG P D,HUANG H,HE L. Improved algorithm for pitch detection based on ACF and CEP[J]. Computer Applications and Software, 2015,32(1):163-166.) [18] GHARAVIAN D,SHEIKHAN M,ASHOFTEDEL F. Emotion recognition improvement using normalized formant supplementary features by hybrid of DTW-MLP-GMM model[J]. Neural Computing and Applications,2013,22(6):1181-1191. [19] NALINI N J,PALANIVEL S,BALASUBRAMANIAN M. Speech emotion recognition using residual phase and MFCC features[J]. International Journal of Engineering and Technology,2014,5(6):4515-4527. [20] 魏云超, 赵耀. 基于DCNN的图像语义分割综述[J]. 北京交通大学学报,2016,40(4):82-91.(WEI Y C,ZHAO Y. A review on image semantic segmentation based on DCNN[J]. Journal of Beijing Jiaotong University,2016,40(4):82-91.) [21] ŚMIEJA M,TABOR J,SPUREK P. SVM with a neutral class[J]. Pattern Analysis and Applications,2019,22(2):573-582. [22] WAN H,GUO S,YIN K,et al. CTS-LSTM:LSTM-based neural networks for correlated time series prediction[J]. KnowledgeBased Systems,2020,191:No. 105239. [23] YUAN T,DENG C,SHI W. Speech emotion recognition based on fuzzy K-NN algorithm with fractionally spaced blind equalization[C]//Proceedings of the 2nd Workshop on Advanced Research and Technology in Industry Applications. Paris:Atlantis Press, 2016:1806-1809. [24] 陈康. 彝语方言研究[M]. 北京:中央民族大学出版社,2010:73-81.(CHEN K. A Study of Yi Dialect[M]. Beijing:China Minzu University Press,2010:73-81.) [25] 庄莉. 彝族的语言使用情况调查[D]. 重庆:四川外国语大学, 2015:47-49.(ZHUANG L. A survey of Yi language use[D]. Chongqing:Sichuan International Studies University, 2015:47-49.) |