Skip to main content

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4682))

Included in the following conference series:

  • 1889 Accesses

Abstract

In learning Bayesian network classifiers, estimating probabilities from a given set of training examples is crucial. In many cases, we can estimate probabilities by the fraction of times the events is observed to occur over the total number of opportunities. However, when the training examples are not enough, this probability estimation method inevitably suffers from the zero-frequency problem. To avoid this practical problem, Laplace estimate is usually used to estimate probabilities. Just as we all know, m-estimate is another probability estimation method. Thus, a natural question is whether a Bayesian network classifier with m-estimate can perform even better. Responding to this question, we single out a special m-estimate method and empirically investigate its effect on various Bayesian network classifiers, such as Naive Bayes (NB), Tree Augmented Naive Bayes (TAN), Averaged One-Dependence Estimators (AODE), and Hidden Naive Bayes (HNB). Our experiments show that the classifiers with our m-estimate perform better than the ones with Laplace estimate.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Chickering, D.M.: Learning Bayesian Networks is NP-Complete. In: Fisher, D., Lenz, H. (eds.) Learning from Data: Artificial Intelligence and Statistics, pp. 121–130. Springer, New York (1996)

    Google Scholar 

  2. Langley, P., Iba, W., Thomas, K.: An Analysis of Bayesian Classifiers. In: Proceedings of the Tenth National Conference of Artificial Intelligence, pp. 223–228. AAAI Press, California, USA (1992)

    Google Scholar 

  3. Friedman, N., Geiger, D., Goldszmidt, M.: Bayesian Network Classifiers. Machine Learning 29, 131–163 (1997)

    Article  MATH  Google Scholar 

  4. Webb, G.I., Boughton, J., Wang, Z.: Not so Naive Bayes: Aggregating One-Dependence Estimators. Machine Learning 58, 5–24 (2005)

    Article  MATH  Google Scholar 

  5. Zhang, H., Jiang, L., Su, J.: Hidden Naive Bayes. In: Proceedings of the 20th National Conference on Artificial Intelligence, pp. 919–924. AAAI Press, California, USA (2005)

    Google Scholar 

  6. Mitchell, T.M.: Machine learning. McGraw-Hill, New York (1997)

    MATH  Google Scholar 

  7. Witten, I.H., Frank, E.: Data Mining: Practical Machine Mearning Tools and Techniques, 2nd edn. Morgan Kaufmann, San Francisco (2005), http://prdownloads.sourceforge.net/weka/datasets-UCI.jar

  8. Merz, C., Murphy, P., Aha, D.: UCI Repository of Machine Learning Databases. In: Dept of ICS, University of California, Irvine (1997), http://www.ics.uci.edu/mlearn/MLRepository.html

  9. Nadeau, C., Bengio, Y.: Inference for the Generalization Error. In: Advances in Neural Information Processing Systems, vol. 12, pp. 307–313. MIT Press, Cambridge (1999)

    Google Scholar 

  10. Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, CA (1993)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

De-Shuang Huang Laurent Heutte Marco Loog

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Jiang, L., Wang, D., Cai, Z. (2007). Scaling Up the Accuracy of Bayesian Network Classifiers by M-Estimate. In: Huang, DS., Heutte, L., Loog, M. (eds) Advanced Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence. ICIC 2007. Lecture Notes in Computer Science(), vol 4682. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74205-0_52

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-74205-0_52

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-74201-2

  • Online ISBN: 978-3-540-74205-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics