Academia.eduAcademia.edu

Outline

K-winner machines for pattern classification

2001, IEEE Transactions on Neural Networks

https://doi.org/10.1109/72.914531

Abstract

The paper describes the K-winner machine (KWM) model for classification. KWM training uses unsupervised vector quantization and subsequent calibration to label data-space partitions. A K-winner classifier seeks the largest set of best-matching prototypes agreeing on a test pattern, and provides a local-level measure of confidence. A theoretical analysis characterizes the growth function of a K-winner classifier, and the result leads to tight bounds to generalization performance. The method proves suitable for high-dimensional multiclass problems with large amounts of data. Experimental results on both a synthetic and a real domain (NIST handwritten numerals) confirm the approach effectiveness and the consistency of the theoretical framework.

References (31)

  1. R. Hecht Nielsen, Neurocomputing. Reading, MA: Addison-Wesley, 1990.
  2. Y. Liu, "Unbiased estimate of generalization error and model selection in neural network," Neural Networks, vol. 8, pp. 215-219, 1995.
  3. L. Prechelt, "Automatic early stopping using cross validation: Quanti- fying the criteria," Neural Networks, vol. 11, no. 4, pp. 761-767, 1998.
  4. H. Akaike, "A new look at the statistical model identification," IEEE Trans. Automat. Contr., vol. AC-19, pp. 716-723, 1974.
  5. J. Rissanen, Stochastic Complexity in Statistical Inquiry, Singapore: World Scientific , 1989.
  6. "Stochastic complexity," J. R. Statist. Soc. B, vol. 49, no. 3, pp. 223-239, 252-265, 1987.
  7. S. Amari, N. Murata, K. R. Mueller, M. Finke, and H. H. Yang, "Asymp- totic statistical theory of overtraining and cross-validation," IEEE Trans. Neural Networks, vol. 8, pp. 985-996, 1997.
  8. V. N. Vapnik, Estimation of Dependences Based on Empirical Data. New York: Springer-Verlag, 1982.
  9. The Nature of Statistical Learning Theory. New York: Springer- Verlag, 1995.
  10. Statistical Learning Theory. New York: Wiley, 1998.
  11. C. J. C. Burgess, "A tutorial on support vector machines for pattern recognition," Data Mining and Knowledge Discovery, vol. 2, no. 2, pp. 1-43, 1998.
  12. T. Kohonen, Self-Organization and Associative Memory, 3rd ed. New York: Springer-Verlag, 1989.
  13. T. Martinetz and K. Schulten, "Topology representing networks," Neural Networks, vol. 7, no. 3, pp. 507-522, 1994.
  14. D. DeSieno, "Adding a conscience to competitive learning," in Proc. IEEE Int. Conf. Neural Networks, vol. 1, San Diego, CA, 1988, pp. 117-124.
  15. T. M. Martinetz, S. G. Berkovich, and K. J. Schulten, ""Neural gas" network for vector quantization and its application to time-series pre- diction," IEEE Trans. Neural Networks, vol. 4, pp. 558-569, 1993.
  16. M. M. Van Hulle, "Kernel-based equiprobabilistic topographic map for- mation," Neural Comput., vol. 10, pp. 1847-1871, 1998.
  17. S. P. Lloyd, "Least squares quantization in PCM," IEEE Trans. Inform. Theory, vol. 28, pp. 127-135, 1982.
  18. N. Ueda and R. Nakano, "A new competitive learning approach based on an equidistortion principle for designing optimal vector quantizers," Neural Networks, vol. 7, no. 8, pp. 1211-1228, 1994.
  19. S. Ridella, S. Rovetta, and R. Zunino, "Plastic algorithm for adaptive vector quantization," Neural Comput. Applicat., vol. 7, no. 1, pp. 37-51, 1998.
  20. S. Rovetta and R. Zunino, "Efficient training of neural gas vector quan- tizers with analog circuit implementation," IEEE Trans. Circuits Syst. II, vol. 46, pp. 688-698, 1999.
  21. J. Weston and C. Watkins, "Multi-class support vector machines," Royal Holloway Univ., Dept. Comp. Sci., London, U.K., CSD-TR-98-04, 1998.
  22. M. Anthony and N. Biggs, Computational Learning Theory. Cambridge, U.K.: Cambridge Univ. Press, 1992.
  23. J. Shawe-Taylor and M. Anthony, "Sample sizes for multiple-output threshold networks," Network Comput. Neural Syst., vol. 2, no. 1, pp. 107-117, Feb. 1991.
  24. D. Haussler, "Generalizing the PAC model: Sample size bounds from metric dimension-based uniform convergence results," in Proc. 30th IEEE Symp. Found. Comput. Sci., 1989, pp. 40-45.
  25. W. Hoeffding, "Probability inequalities for sums of bounded random variables," J. Amer. Stat. Assoc., vol. 58, pp. 13-30, 1963.
  26. A. F. R. Rahman and M. C. Fairhurst, "An evaluation of multiexpert con- figurations for the recognition of handwritten numerals," Pattern Recog- nition, vol. 31, no. 9, pp. 1255-1273, 1998.
  27. A. M. Colla and P. Pedrazzi, "Binary digital image feature extracting process," U.S. Pat. no. 5 737 444, Apr. 7, 1998.
  28. S. Zunino, "Circular backpropagation networks embed Vector quantiza- tion," IEEE Trans. Neural Networks, vol. 10, pp. 972-975, July 1999. [29] , "Circuit implementation of the K-Winner Machine," Electron.
  29. Lett., vol. 35, no. 14, pp. 1172-1173, July 8, 1999.
  30. Sandro Ridella (M'93) received the Laurea degree in electronic engineering from the University of Genova, Italy, in 1966. He is a full Professor in the Department of Biophysical and Electronic Engi- neering, University of Genova, Italy, where he teaches Circuits and Algorithms for Signal Processing. In the last eight years, his scientific activity has been mainly focused in the field of neural networks.
  31. Stefano Rovetta (M'98) received the Laurea degree in electronic engineering and the Ph.D. degree in models, methods, and tools for electronic and electro- magnetic systems, both from the University of Genoa, Italy, in 1993 and 1997, respectively. He held a position as a postdoctoral researcher with the Electronic Systems and Networking Group of the Department of Biophysical and Electronic Engi- neering, University of Genoa. He has been an Invited Professor of Operating Systems at the University of Siena. He is currently an Assistant Professor at the Department of Computer and Information Sciences of Genoa University. His research interests include electronic circuits and systems, and neural-network theory, implementation, and applications. Rodolfo Zunino (S'90-M'90) received the Laurea degree in electronic engineering from Genoa Univer- sity, Italy, in 1985. From 1986 to 1995, he was a Research Consul- tant with the Department of Biophysical and Elec- tronic Engineering of Genoa University. He is cur- rently with the same department as an Associate Pro- fessor in Electronics and Industrial Electronics. His main scientific interests include electronic systems for neural networks, efficient models for data repre- sentation and learning, advanced techniques for mul- timedia data processing, and distributed-control methodologies.