Academia.eduAcademia.edu

Outline

Classification Methods

Encyclopedia of Data Warehousing and Mining, Second Edition

https://doi.org/10.4018/978-1-60566-010-3.CH032

Abstract

Generally speaking, classification is the action of assigning an object to a category according to the characteristics of the object. In data mining, classification refers to the task of analyzing a set of pre-classified data objects to learn a model (or a function) that can be used to classify an unseen data object into one of several predefined classes. A data object, referred to as an example, is described by a set of attributes or variables. One of the attributes describes the class that an example belongs to and is thus called the class attribute or class variable. Other attributes are often called independent or predictor attributes (or variables). The set of examples used to learn the classification model is called the training data set. Tasks related to classification include regression, which builds a model from training data to predict numerical values, and clustering, which groups examples to form categories. Classification belongs to the category of supervised learning, ...

References (20)

  1. An, A., & Cercone, N. (1998). ELEM2: A learning system for more accurate classifications. Proceedings of the 12th Canadian Conference on Artificial Intelli- gence, 426-441.
  2. Breiman, L., Friedman, J., Olshen, R., & Stone, C. (1984). Classification and regression trees, Wadsworth International Group.
  3. Buntine, W.L. (1994). Operations for learning with graphical models. Journal of Artificial Intelligence Research, 2, 159-225.
  4. Castillo, E., Gutiérrez, J.M., & Hadi, A.S. (1997). Ex- pert systems and probabilistic network models. New York: Springer-Verlag.
  5. Clark P., & Boswell, R. (1991). Rule induction with CN2: Some recent improvements. Proceedings of the 5 th European Working Session on Learning, 151-163.
  6. Cohen, W.W. (1995). Fast effective rule induction. Proceedings of the 11th International Conference on Machine Learning, 115-123, Morgan Kaufmann.
  7. Cooper, G., & Herskovits, E. (1992). A Bayesian method for the induction of probabilistic networks from data. Machine Learning, 9, 309-347.
  8. Domingos, P., & Pazzani, M. (1997). On the optimality of the simple Bayesian classifier under zero-one loss. Machine Learning, 29, 103-130.
  9. Gehrke, J., Ramakrishnan, R., & Ganti, V. (1998). RainForest -A framework for fast decision tree con- struction of large datasets. Proceedings of the 24 th International Conference on Very Large Data Bases.
  10. Han, J., & Kamber, M. (2001). Data mining -Con- cepts and techniques. Morgan Kaufmann.
  11. Heckerman, D., Geiger, D., & Chickering, D.M. (1995) Learning bayesian networks: The combination of knowl- edge and statistical data. Machine Learning, 20, 197-243.
  12. Mitchell, T.M. (1997). Machine learning. McGraw-Hill.
  13. Nurnberger, A., Pedrycz, W., & Kruse, R. (2002). Neu- ral network approaches. In Klosgen & Zytkow (Eds.), Handbook of data mining and knowledge discovery. Oxford University Press.
  14. Pearl, J. (1986). Fusion, propagation, and structuring in belief networks. Artificial Intelligence, 29(3), 241-288.
  15. Quinlan, J.R. (1993). C4.5: Programs for machine learn- ing. Morgan Kaufmann.
  16. Rish, I. (2001). An empirical study of the naive Bayes classifier. Proceedings of IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence.
  17. Rumelhart, D.E., Hinton, G.E., & Williams, R.J. (1986). Learning representations by back-propagating errors. Nature, 323, 533-536.
  18. Russell, S., Binder, J., Koller, D., & Kanazawa, K. (1995). Local learning in probabilistic networks with hidden variables. Proceedings of the 14 th Joint International Conference on Artificial Intelligence, 2, 1146-1152.
  19. Shafer, J., Agrawal, R., & Mehta, M. (1996). SPRINT: A scalable parallel classifier for data mining. Proceed- ings of the 22 th International Conference on Very Large Data Bases.
  20. Vapnik, V. (1998). Statistical learning theory. New York: John Wiley & Sons.