Abstract:Locally Weighted Naive Bayes (LWNB) is a good improvement of Naive Bayes (NB) and Discriminative Frequency Estimate (DFE) remarkably improves the generalization accuracy of Naive Bayes. Inspired by LWNB and DFE, this paper proposed Gradually Contracting Spaces (GCS) algorithm to learn parameters of Naive Bayes. Given a test instance, GCS found a series of subspaces in global space which contained all training instances. All of these subspaces contained the test instance and any of them must be contained by others that are bigger than it. Then GCS used training instances contained in those subspaces to gradually learn parameters of Naive Bayes (NB) by Modified version of DFE (MDFE) which was a modified version of DFE and used NB to classify test instances. GSC trained Naive Bayes with all training data and achieved an eager version, which was the essential difference between GSC and LWNB. Decision tree version of GCS named GCS-T was implemented in this paper. The experimental results show that GCS-T has higher generalization accuracy compared with C4.5 and some Bayesian classification algorithms such as Naive Bayes, BaysianNet, NBTree, Hidden Naive Bayes (HNB), LWNB, and the classification speed of GCS-T is remarkably faster than LWNB.
THEODORIDIS S, KOUTROUMBAS K. Pattern recognition [M]. 4th ed. Maryland Heights, MO: Elsevier, 2009.
[2]
KOHAVI R. Scaling up the accuracy of nave Bayes classifiers: A decision-tree hybrid [C]// Proceedings of the Second International Conference on Knowledge Discovery and Data Mining. New York: ACM Press, 1996: 202-207.
[3]
FRIEDMAN N, GEIGER D, GOLDSZMIDT M. Bayesian network classifiers [J]. Machine Learning, 1997, 29(2/3): 131-163.
[4]
张连文,郭海鹏.贝叶斯网络引论[M].北京:科学出版社,2006.
[5]
FRANK E, HALL M, PFAHRINGER B. Locally weighted naive Bayes [C]// Proceedings of the 19th Conference in Uncertainty in Artificial Intelligence. Seattle: Morgan Kaufmann, 2003: 249-256.
[6]
JIANG L, ZHANG H, CAI Z. A novel Bayes model: Hidden naive Bayes [J]. IEEE Transactions on Knowledge and Data Engineering, 2009, 21(10): 1361-1371.
[7]
KAI M T JONATHAN R W, SWEE C T, et al. Feature-subspace aggregating: Ensembles for stable and unstable learners [J]. Machine Learning, 2011, 82(3): 375-397.
[8]
PERNKOPF F, WOHLMAYR M. On discriminative parameter learning of Bayesian network classifiers [C]// European Conference on Machine Learning and Principle and Practice of Knowledge Discovery in Databases. Berlin: Springer-Verlag, 2009: 221-237.
[9]
GREINER R, SU X, SHEN B, et al. Structural extension to logistic regression: Discriminative parameter learning of belief net classifiers [J]. Machine Learning, 2005, 59(3): 297-322.
[10]
SU J, ZHANG H, LING C X, et al. Discriminative parameter learning for Bayesian networks [C]// Proceedings of the 25th International Conference on Machine Learning. New York: ACM Press, 2008: 1016-1023.
[11]
QUINLAN J R. C4.5: Programs for machine learning [M]. Seattle: Morgan Kaufmann, 1993.
[12]
ASUNCION A, NEWMAN D J. UCI repository of machine learning databases [EB/OL]. [2011-03-25]. http://www.ics.uci.edu/~mlearn/ MLRepository.html.
[13]
WITTEN I H, FRANK E, HALL M A. Data mining: Practical machine learning tools and techniques [M]. 3rd ed. Seattle: Morgan Kaufmann, 2011.