Academia.eduAcademia.edu

Outline

Reduction Techniques for Instance-Based Learning Algorithms

https://doi.org/10.1023/A:1007626913721

Abstract

Instance-based learning algorithms are often faced with the problem of deciding which instances to store for use during generalization. Storing too many instances can result in large memory requirements and slow execution speed, and can cause an oversensitivity to noise. This paper has two main purposes. First, it provides a survey of existing algorithms used to reduce storage requirements in instance-based learning algorithms and other exemplar-based algorithms. Second, it proposes six additional reduction algorithms called DROP1{DROP5 and DEL (three of which were rst described in Wilson & Martinez, 1997c, as RT1{RT3) that can be used to prune instances from the concept description. These algorithms and 10 algorithms from the survey are compared on 31 classi cation tasks. Of those algorithms that provide substantial storage reduction, the DROP algorithms have the highest average generalization accuracy in these experiments, especially in the presence of noise.

References (49)

  1. Aha, David W. (1992). Tolerating noisy, irrelevant a n d n o vel attributes in instance-based learning algorithms. International Journal of Man-Machine Studies, 36, pp. 267{287.
  2. Aha, David W., Dennis Kibler, Marc K. Albert (1991). Instance-Based Learning Algorithms. Machine Learning, 6, pp. 37{66.
  3. Batchelor, Bruce G. (1978). Pattern Recognition: Ideas in Practice. New York: Plenum Press.
  4. Biberman, Yoram (1994). A C o n text Similarity Measure. In Proceedings of the European Con- ference on Machine Learning (ECML-94). Catania, Italy: Springer Verlag, pp. 49{63.
  5. Brodley, Carla E. (1993). Addressing the Selective Superiority Problem: Automatic Algorithm/Model Class Selection. Proceedings of the Tenth International Machine Learning Conference, Amherst, MA, pp. 17{24.
  6. Broomhead, D. S., and D. Lowe (1988). Multi-variable functional interpolation and adaptive networks. Complex Systems, 2, pp. 321{355.
  7. Cameron-Jones, R. M. (1995). Instance Selection by Encoding Length Heuristic with Random Mutation Hill Climbing. In Proceedings of the Eighth Australian Joint Conference o n A rti cial Intelligence, pp. 99{106.
  8. Carpenter, Gail A., and Stephen Grossberg (1987). A Massively Parallel Architecture for a Self- Organizing Neural Pattern Recognition Machine. Computer Vision, Graphics, and Image Pro- cessing, 37, pp. 54{115.
  9. Chang, Chin-Liang (1974). Finding Prototypes for Nearest Neighbor Classi ers. IEEE Transac- tions on Computers, 23{11, N o vember 1974, pp. 1179{1184.
  10. Conover, W. J. (1971). Practical Nonparametric Statistics. N e w Y ork: John Wiley, pp. 206{209, 383. Cover, T. M., and P. E. Hart (1967). Nearest Neighbor Pattern Classi cation. Institute of Electrical and Electronics Engineers Transactions on Information Theory, 13{1, J a n uary 1967, pp. 21{27.
  11. Dasarathy, Belur V., (1991). Nearest Neighbor (NN) Norms: NN Pattern Classi cation Tech- niques, Los Alamitos, CA: IEEE Computer Society P r e s s .
  12. DeGroot, M. H. (1986). Probability and Statistics (Second Edition). Reading, MA: Addison- Wesley.
  13. Diday, Edwin (1974). Recent Progress in Distance and Similarity Measures in Pattern Recognition. Second International Joint Conference on Pattern Recognition, pp. 534{539.
  14. Dietterich, Thomas G. (1989). Limitations on Inductive Learning. In Proceedings of the Sixth International Conference on Machine Learning. San Mateo, CA: Morgan Kaufmann, pp. 124{ 128. Domingos, Pedro (1995). Rule Induction and Instance-Based Learning: A Uni ed Approach. In Proceedings of the Fourteenth International Joint Conference o n A rti cial Intelligence (IJCAI- 95), Montreal, Canada: Morgan Kaufmann, pp. 1226{1232.
  15. Domingos, Pedro (1996). Unifying Instance-Based and Rule-Based Induction. Machine Learning, 24, pp. 141{168.
  16. Dudani, Sahibsingh A. (1976). The Distance-Weighted k-Nearest-NeighborRule. IEEE Transac- tions on Systems, Man and Cybernetics, 6{4, April 1976, pp. 325{327.
  17. Gates, G. W. (1972). The Reduced Nearest Neighbor Rule. IEEE Transactions on Information Theory, IT{18{3, pp. 431{433.
  18. Hart, P. E . ( 1 9 6 8 ) . The Condensed Nearest Neighbor Rule. IEEE Transactions on Information Theory, 14, pp. 515{516.
  19. Hecht-Nielsen, R. (1987). Counterpropagation Networks. Applied Optics, 26{23, p p . 4979{4984.
  20. Kibler, D., and David W. Aha (1987). Learning Representative Exemplars of Concepts: An Initial Case Study. Proceedings of the Fourth International Workshop on Machine Learning, Irvine, CA: Morgan Kaufmann, pp. 24{30.
  21. Lowe, David G. (1995). Similarity Metric Learning for a Variable-Kernel Classi er. Neural Com- putation. 7{1, pp. 72{85.
  22. Merz, C. J., and P. M . M u r h y (1996). UCI Repository of Machine Learning Databases. Irvine, CA: University of California Irvine, Department of Information and Computer Science. Internet: http://www.ics.uci.edu/ ~mlearn/ MLRepository.html.
  23. Michalski, Ryszard S., Robert E. Stepp, and Edwin Diday (1981). A Recent A d v ance in Data Analysis: Clustering Objects into Classes Characterized by Conjunctive Concepts. Progress in Pattern Recognition, 1, L a veen N. Kanal and Azriel Rosenfeld (Eds.). New York: North-Holland, pp. 33{56.
  24. Mitchell, Tom M., (1980). The Need for Biases in Learning Generalizations. In J. W. Shavlik & T. G. Dietterich (Eds.), Readings in Machine Learning, San Mateo, CA: Morgan Kaufmann, 1990, pp. 184{191.
  25. Nadler, Morton, and Eric P. Smith (1993). Pattern Recognition Engineering. N e w Y ork: Wiley.
  26. Papadimitriou, C. H., and Steiglitz, K. (1982). Combinatorial Optimization: Algorithms and Complexity. Prentice-Hall, Englewood Cli s, NJ.
  27. Papadimitriou, Christos H., and Jon Louis Bentley (1980). A W orst-Case Analysis of Nearest Neighbor Searching by Projection. Lecture Notes in Computer Science, 85, Automata Languages and Programming, pp. 470{482.
  28. Renals, Steve, and Richard Rohwer (1989). Phoneme Classi cation Experiments Using Radial Basis Functions. In Proceedings of the IEEE International Joint Conference o n N e u r al Networks (IJCNN'89), 1, pp. 461{467.
  29. Ritter, G. L., H. B. Woodru , S. R. Lowry, and T. L. Isenhour (1975). An Algorithm for a Selective Nearest Neighbor Decision Rule. IEEE Transactions on Information Theory, 21{6, November 1975, pp. 665{669.
  30. Rumelhart, D. E., and J. L. McClelland (1986). Parallel Distributed P r ocessing, MIT Press, Ch. 8, pp. 318{362.
  31. Salzberg, Steven (1991). A Nearest Hyperrectangle Learning Method. Machine Learning, 6, pp. 277{309. Scha er, Cullen (1994). A Conservation Law for Generalization Performance. In Proceedings of the Eleventh International Conference on Machine Learning (ML'94), New Brunswick, NJ: Morgan Kaufmann, pp. 259{265.
  32. Skalak, D. B. (1994). Prototype and Feature Selection by Sampling and Random Mutation Hill Climbing Algorithms. In Proceedings of the Eleventh International Conference on Machine Learning (ML94). Morgan Kaufmann, pp. 293{301.
  33. Specht, Donald F. (1992). Enhancements to Probabilistic Neural Networks. In Proceedings In- ternational Joint Conference on Neural Networks (IJCNN '92), 1, pp. 761{768.
  34. Sproull, Robert F. (1991). Re nements to Nearest-Neighbor Searching in k-Dimensional Trees. Algorithmica, 6, pp. 579{589.
  35. Stan ll, C., and D. Waltz (1986). Toward memory-based reasoning. Communications of the ACM, 29, pp. 1213{1228.
  36. Tomek, Ivan (1976). An Experiment with the Edited Nearest-Neighbor Rule. IEEE Transactions on Systems, Man, and Cybernetics, 6{6, pp. 448{452.
  37. Tversky, Amos (1977). Features of Similarity. Psychological Review, 84{4, pp. 327{352.
  38. Wasserman, Philip D. (1993). Advanced Methods in Neural Computing. New York, NY: Van Nostrand Reinhold, pp. 147{176.
  39. Watson, I., and F. Marir (1994). Case-Based Reasoning: A Review. The Knowledge Engineering Review, 9{4, C a m bridge, UK: Cambridge University P r e s s .
  40. Wess, Stefan, Klaus-Dieter Altho , and Michael M. Richter (1993), Using k-d Trees to Improve the Retrieval Step in Case-Based Reasoning. Topics in Case-Based R easoning, First European Workshop (EWCBR-93), Springer-Verlag, pp. 67{181.
  41. Wettschereck, Dietrich (1994). A Hybrid Nearest-Neighbor and Nearest-Hyperrectangle Algo- rithm, In Proceedings of the 7th European Conference on Machine Learning, LNAI-784, F. Bergadano and L. de Raedt (editors), pp. 323{335.
  42. Wettschereck, Dietrich, and Thomas G. Dietterich (1995). An Experimental Comparison of Nearest-Neighbor and Nearest-Hyperrectangle Algorithms. Machine Learning, 19{1, pp. 5{ 28. Wettschereck, Dietrich, David W. Aha, and Takao Mohri (1997). A Review and Comparative Evaluation of Feature Weighting Methods for a Class of Lazy Learning Algorithms. Arti cial Intelligence R eview, 11, issues Special Issue on Lazy Learning, Klewer Academic Publishers, pp. 273{314.
  43. Wilson, D. Randall, and Tony R. Martinez (1996). Heterogeneous Radial Basis Functions. Pro- ceedings of the International Conference on Neural Networks (ICNN'96), 2, pp. 1263{1267.
  44. Wilson, D. Randall, and Tony R. Martinez (1997a). Improved Heterogeneous Distance Functions. Journal of Arti cial Intelligence R esearch (JAIR), 6{1, pp. 1{34.
  45. Wilson, D. Randall, and Tony R. Martinez (1997b). Improved Center Point Selection for Radial Basis Function Networks. In Proceedings of the International Conference o n A rti cial Neural Networks and Genetic Algorithms (ICANNGA'97).
  46. Wilson, D. Randall, and Tony R. Martinez (1997c). Instance Pruning Techniques. In Fisher, D., ed., Machine Learning: Proceedings of the Fourteenth International Conference (ICML'97), Morgan Kaufmann Publishers, San Francisco, CA, pp. 403{411.
  47. Wilson, Dennis L. (1972). Asymptotic Properties of Nearest Neighbor Rules Using Edited Data. IEEE Transactions on Systems, Man, and Cybernetics, 2{3, pp. 408{421.
  48. Wolpert, David H. (1993). On Over tting Avoidance as Bias. Technical Report SFI TR 92-03- 5001. Santa Fe, NM: The Santa Fe Institute.
  49. Zhang, Jianping (1992). Selecting Typical Instances in Instance-Based Learning. Proceedings of the Ninth International Conference on Machine Learning, Aberdeen, Scotland: Morgan Kauf- mann, pp. 470{479.