计算机应用 ›› 2016, Vol. 36 ›› Issue (2): 291-294.DOI: 10.11772/j.issn.1001-9081.2016.02.0291

• 第三届CCF大数据学术会议(CCF BigData 2015) •    下一篇

基于概率校准的集成学习

姜正申1, 刘宏志2   

  1. 1. 北京大学 信息科学技术学院, 北京 100871;
    2. 北京大学 软件与微电子学院, 北京 102600
  • 收稿日期:2015-08-29 修回日期:2015-09-11 出版日期:2016-02-10 发布日期:2016-02-03
  • 通讯作者: 刘宏志(1982-),男,湖北武汉人,副教授,博士,CCF会员,主要研究方向:信息融合、模式识别。
  • 作者简介:姜正申(1990-),男,吉林松原人,博士研究生,主要研究方向:模式识别。
  • 基金资助:
    国家自然科学基金资助项目(61232005);CCF-腾讯科研基金资助项目。

Ensemble learning based on probability calibration

JIANG Zhengshen1, LIU Hongzhi2   

  1. 1. School of Electronics Engineering and Computer Science, Peking University, Beijing 100871, China;
    2. School of Software and Microelectronics, Peking University, Beijing 102600, China
  • Received:2015-08-29 Revised:2015-09-11 Online:2016-02-10 Published:2016-02-03

摘要: 针对原有集成学习多样性不足而导致的集成效果不够显著的问题,提出一种基于概率校准的集成学习方法以及两种降低多重共线性影响的方法。首先,通过使用不同的概率校准方法对原始分类器给出的概率进行校准;然后使用前一步生成的若干校准后的概率进行学习,从而预测最终结果。第一步中使用的不同概率校准方法为第二步的集成学习提供了更强的多样性。接下来,针对校准概率与原始概率之间的多重共线性问题,提出了选择最优(choose-best)和有放回抽样(bootstrap)的方法。选择最优方法对每个基分类器,从原始分类器和若干校准分类器之间选择最优的进行集成;有放回抽样方法则从整个基分类器集合中进行有放回的抽样,然后对抽样出来的分类器进行集成。实验表明,简单的概率校准集成学习对学习效果的提高有限,而使用了选择最优和有放回抽样方法后,学习效果得到了较大的提高。此结果说明,概率校准为集成学习提供了更强的多样性,其伴随的多重共线性问题可以通过抽样等方法有效地解决。

关键词: 集成学习, 概率校准, 多重共线性, 有放回抽样, 随机子空间

Abstract: Since the lackness of diversity may lead to bad performance in ensemble learning, a new two-phase ensemble learning method based on probability calibration was proposed, as well as two methods to reduce the impact of multiple collinearity. In the first phase, the probabilities given by the original classifiers were calibrated using different calibration methods. In the second phase, another classifier was trained using the calibrated probabilities and the final result was predicted. The different calibration methods used in the first phase provided diversity for the second phase, which has been shown to be an important factor to enhance ensemble learning. In order to address the limited improvement due to the correlation between base classifiers, two methods to reduce the multiple collinearity were also proposed, that is, choose-best and bootstrap sampling method. The choose-best method just selected the best base classifier among original and calibrated classifiers; the bootstrap method combined a set of classifiers, which were chosen from the base classifiers with replacement. The experimental results showed that the use of different calibrated probabilities indeed improved the effectiveness of the ensemble; after using the choose-best and bootstrap sampling methods, further improvement was also achieved. It means that probability calibration provides a new way to produce diversity, and the multiple collinearity caused by it can be solved by sampling method.

Key words: ensemble learning, probability calibration, multiple collinearity, bootstrap sampling, random subspace

中图分类号: