计算机应用 ›› 2016, Vol. 36 ›› Issue (12): 3333-3340.DOI: 10.11772/j.issn.1001-9081.2016.12.3333

• 人工智能 • 上一篇    下一篇

基于深度卷积神经网络的物体识别算法

黄斌, 卢金金, 王建华, 吴星明, 陈伟海   

  1. 北京航空航天大学 自动化科学与电气工程学院, 北京 100191
  • 收稿日期:2016-04-26 修回日期:2016-07-11 出版日期:2016-12-10 发布日期:2016-12-08
  • 通讯作者: 陈伟海
  • 作者简介:黄斌(1989-),男,安徽安庆人,博士研究生,主要研究方向:计算机视觉、深度学习、机器学习;卢金金(1990-),女,安徽安庆人,硕士研究生,主要研究方向:计算机视觉、深度学习、机器学习;王建华(1962-),男,北京人,副教授,博士,主要研究方向:智能机器人检测和控制、计算机视觉;吴星明(1962-),男,北京人,副教授,博士,主要研究方向:智能机器人检测和控制、计算机视觉;陈伟海(1955-),男,浙江象山人,教授,博士,主要研究方向:康复机器人、计算机视觉。
  • 基金资助:
    国家自然科学基金资助项目(61573048);重点国际(地区)合作研究项目(61620106012);国家国际科技合作专项(2015DFG12650)。

Object recognition algorithm based on deep convolution neural networks

HUANG Bin, LU Jinjin, WANG Jianhua, WU Xingming, CHEN Weihai   

  1. School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
  • Received:2016-04-26 Revised:2016-07-11 Online:2016-12-10 Published:2016-12-08
  • Supported by:
    This work is partially supported by the National Natural Science Foundation of China (61573048), the Major International (Regional) Joint Research Project (61620106012), the International Scientific and Technological Cooperation Projects of China under Grant (2015DFG12650).

摘要: 针对传统物体识别算法中人工设计出来的特征易受物体形态多样性、光照和背景的影响,提出了一种基于深度卷神经网络的物体识别算法。该算法基于NYU Depth V2场景数据库,首先将单通道深度信息转换为三通道;再用训练集中的彩色图片和转换后的三通道深度图片分别微调两个深度卷积神经网络模型;然后用训练好的模型对重采样训练集中的彩色和深度图片提取模型第一个全连接层的特征,并将两种模态的特征串联起来,训练线性支持向量机(LinSVM);最后将所提算法应用到场景理解任务中的超像素特征提取。所提方法在测试集上的物体分类准确度可达到91.4%,比SAE-RNN方法提高4.1个百分点。实验结果表明所提方法可提取彩色和深度图片高层特征,有效提高物体分类准确度。

关键词: 计算机视觉, 卷积神经网络, 特征提取, 线性支持向量机, 物体识别, 场景理解

Abstract: Focused on the problem of traditional object recognition algorithm that the artificially designed features were more susceptible to diversity of object shapes, illumination and background, a deep convolutional neural network algorithm was proposed for object recognition. Firstly, this algorithm was trained with NYU Depth V2 dataset, and single depth information was transformed into three channels. Then color images and transformed depth images in the training set were used to fine-tune two deep convolutional neural networks, respectively. Next, color and depth image features were extracted from the first fully connected layers of the two trained models, and the two features from the resampling training set were combined to train a Linear Support Vector Machine (LinSVM) classifier. Finally, the proposed object recognition algorithm was used to extract super-pixel features in scene understanding task. The proposed method can achieve a classification accuracy of 91.4% on the test set which is 4.1 percentage points higher than SAE-RNN (Sparse Auto-Encoder with the Recursive Neural Networks). The experimental results show that the proposed method is effective in extracting color and depth image features, and can effectively improve classification accuracy.

Key words: computer vision, Convolutional Neural Network (CNN), feature extraction, Linear Support Vector Machine (LinSVM), object recognition, scene understanding

中图分类号: