[1] 徐超,闫胜业.改进的卷积神经网络行人检测方法[J].计算机应用,2017,37(6):1708-1715.(XU C, YAN S Y. Improved pedestrian detection method based on convolutional neural network[J]. Journal of Computer Applications, 2017, 37(6):1708-1715.) [2] HINTERSTOISSER S, LEPETIT V, ILIC S, et al. Model based training, detection and pose estimation of texture-less 3D objects in heavily cluttered scenes[C]//ACCV 2012:Proceedings of the 11th Asian Conference on Computer Vision. Berlin:Springer, 2012:548-562. [3] SCHUSTER S, KRISHNA R, CHANG A, et al. Generating semantically precise scene graphs from textual descriptions for improved image retrieval[C]//Proceedings of the 2015 Fourth Workshop on Vision and Language. Stroudsburg, PA:ACL, 2015:70-80. [4] SHOTTON J, WINN J, ROTHER C, et al. Textonboost for image understanding:multi-class object recognition and segmentation by jointly modeling texture, layout, and context[J]. International Journal of Computer Vision, 2009, 81(1):2-23. [5] FARABET C, COUPRIE C, NAJMAN L, et al. Learning hierarchical features for scene labeling[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2013, 35(8):1915-1929. [6] LONG J, SHELHAMER E, DARRELL T. Fully convolutional networks for semantic segmentation[C]//Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ:IEEE, 2015:3431-3440. [7] KHAN S H, BENNAMOUN M, SOHEL F, et al. Integrating geometrical context for semantic labeling of indoor scenes using RGBD images[J]. International Journal of Computer Vision, 2016, 117(1):1-20. [8] GUPTA S, ARBELÁEZ P, GIRSHICK R, et al. Indoor scene understanding with RGB-D images:bottom-up segmentation, object detection and semantic segmentation[J]. International Journal of Computer Vision, 2015, 112(2):133-149. [9] KENDALL A, BADRINARAYANAN V, CIPOLLA R. Bayesian SegNet:model uncertainty in deep convolutional encoder-decoder architectures for scene understanding[EB/OL].[2017-04-10]. http://pdfs.semanticscholar.org/9694/c4d214a59979ee182136e9dfb2975dfebaa2.pdf. [10] REN X F, BO L F, FOX D. RGB-(D) scene labeling:features and algorithms[C]//Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition. Washington, DC:IEEE Computer Society, 2012:2759-2766. [11] WANG A, LU J, CAI J, et al. Unsupervised joint feature learning and encoding for RGB-D scene labeling[J]. IEEE Transactions on Image Processing, 2015, 24(11):4459-4473. [12] LI Z, GAN Y K, LIANG X D, et al. LSTM-CF:unifying context modeling and fusion with LSTMs for RGB-D scene labeling[C]//Proceedings of the 2016 European Conference on Computer Vision, LNCS 9906. Berlin:Springer, 2016:541-557. [13] CADENA C, KOŠECKÁ J. Semantic segmentation with heterogeneous sensor coverages[C]//Proceedings of the 2014 IEEE Conference on Robotics and Automation, Washington, DC:IEEE Computer Society, 2014:2639-2645. [14] ZHENG S, JAYASUMANA S, ROMERA-PAREDES B, et al. Conditional random fields as recurrent neural networks[EB/OL].[2017-04-10]. http://www.robots.ox.ac.uk/~szheng/papers/CRFasRNN.pdf. [15] LIN G S, SHEN C H, VAN DEN HENGEL A, et al. Efficient piecewise training of deep structured models for semantic segmentation[C]//Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition. Washington, DC:IEEE Computer Society, 2016:3194-3203. [16] 李艳丽,周忠,吴威.一种双层条件随机场的场景解析方法[J].计算机学报,2013,36(9):1898-1907.(LI Y L, ZHOU Z, WU W. Scene parsing based on a two-level conditional random field[J]. Chinese Journal of Computers, 2013, 36(9):1898-1907.) [17] LIANG X D, SHEN X H, FENG J S, et al. Semantic object parsing with graph LSTM[C]//Proceedings of the 2016 European Conference on Computer Vision, LNCS 9905. Berlin:Springer, 2016:125-143. [18] SILBERMAN N, HOIEM D, KOHLI P, et al. Indoor segmentation and support inference from RGBD images[C]//Proceedings of the 201212th European Conference on Computer Vision, LNCS 7576. Berlin:Springer, 2012:746-760. [19] SONG S, LICHTENBERG S P, XIAO J X. SUN RGB-D:a RGB-D scene understanding benchmark suite[C]//Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ:IEEE, 2015:567-576. [20] GUPTA S, GIRSHICK R, ARBELÁEZ P, et al. Learning rich features from RGB-D images for object detection and segmentation[C]//Proceedings of the 2014 European Conference on Computer Vision, LNCS 8695. Berlin:Springer, 2014:345-360. [21] ACHANTA R, SHAJI A, SMITH K, et al. SLIC superpixels compared to state-of-the-art superpixel methods[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, 34(11):2274-2282. [22] 王春波,董红斌,印桂生,等.基于Hadoop的超像素分割算法[J].计算机应用,2016,36(11):2985-2992.(WANG C B, DONG H B, YIN G S, et al. Super pixel segmentation algorithm based on Hadoop[J]. Journal of Computer Applications, 2016, 36(11):2985-2992.) [23] SMITH T, GUILD J. The CIE colorimetric standards and their use[J]. Transactions of the Optical Society, 1931, 33(3):73-134. [24] HINTON G E, SALAKHUTDINOV R R. Reducing the dimensionality of data with neural networks[J]. Science, 2006, 313(5786):504-507. [25] FREUND Y, HAUSSLER D. Unsupervised learning of distributions on binary vectors using two layer networks, Technical Report UCSC-CRL-94-25[R]. Santa Cruz, CA:University of California at Santa Cruz, 1994. [26] HINTON G E, OSINDERO S, TEH Y W. A fast learning algorithm for deep belief nets[J]. Neural Computation, 2006, 18(7):1527-1554. [27] HINTON G E. Training products of experts by minimizing contrastive divergence[J]. Neural Computation, 2002, 14(8):1771-1800. [28] SCHMIDT M. UGM:a matlab toolbox for probabilistic undirected graphical models[EB/OL].[2017-04-10]. http://www.cs.ubc.ca/~schmidtm/Software/UGM.html. [29] PERCEPTRON M. DeepLearning 0.1 documentation[EB/OL].[2017-04-10]. http://deeplearning.net/tutorial/. [30] JANOCH A, KARAYEV S, JIA Y Q, et al. A category-level 3D object dataset:putting the kinect to work[M]//Consumer Depth Cameras for Computer Vision. London:Springer, 2013:141-165. [31] XIAO J X, OWENS A, TORRALBA A. SUN3D:a database of big spaces reconstructed using sfm and object labels[C]//Proceedings of the 2013 IEEE International Conference on Computer Vision. Washington, DC:IEEE Computer Society, 2013:1625-1632. |