Academia.eduAcademia.edu

Outline

Adaptive Distance Metric Learning for Clustering

2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition

https://doi.org/10.1109/CVPR.2007.383103

Abstract

A good distance metric is crucial for unsupervised learning from high-dimensional data. To learn a metric without any constraint or class label information, most unsupervised metric learning algorithms appeal to projecting observed data onto a low-dimensional manifold, where geometric relationships such as local or global pairwise distances are preserved. However, the projection may not necessarily improve the separability of the data, which is the desirable outcome of clustering. In this paper, we propose a novel unsupervised Adaptive Metric Learning algorithm, called AML, which performs clustering and distance metric learning simultaneously. AML projects the data onto a low-dimensional manifold, where the separability of the data is maximized. We show that the joint clustering and distance metric learning can be formulated as a trace maximization problem, which can be solved via an iterative procedure in the EM framework. Experimental results on a collection of benchmark data sets demonstrated the effectiveness of the proposed algorithm.

References (28)

  1. M. Belkin, I. Matveeva, and P. Niyogi. Regulariza- tion and semi-supervised learning on large graphs. In COLT, pages 624-638, 2004. 6
  2. M. Belkin and P. Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. In NIPS, 2003. 1
  3. C. Blake and C. Merz. UCI repository of machine learning databases, 1998. 2
  4. O. Chapelle, B. Schölkopf, and A. Zien. Semi- Supervised Learning. The MIT Press, 2006. 2, 5
  5. O. Delalleau, Y. Bengio, and N. L. Roux. Efficient non-parametric function induction in semi-supervised learning. In AISTATS, 2005. 6
  6. I. S. Dhillon, Y. Guan, and B. Kulis. A unified view of kernel k-means, spectral clustering and graph par- titioning. Technical report, Department of Computer Sciences, University of Texas at Austin, 2005. 3
  7. P. Diaconis and D. Freedman. Asymptotics of graphical projection pursuit. Annuals of Statistics, 12:793815, 1984. 2
  8. C. Ding, X. He, and H. D. Simon. On the equivalence of nonnegative matrix factorization and spectral clus- tering. In SDM, pages 606-610, 2006. 2, 3
  9. J. H. Friedman. Regularized discriminant analy- sis. Journal of the American Statistical Association, 84(405):165-175, 1989. 2
  10. A. Georghiades, P. Belhumeur, and D. Kriegman. From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intelligence, 23:643-660, 2001. 2
  11. P. Hall and K. Li. On almost linearity of low dimen- sional projections from high dimensional data. Annu- als of Statistics, 21:867-889, 1993. 2
  12. T. Hastie, R. Tibshirani, and J. Friedman. The Ele- ments of Statistical Learning: Data Mining, Inference, and Prediction. Springer, 2001. 3
  13. X. He, D. Cai, and P. Niyogi. Laplacian score for fea- ture selection. In NIPS. 4
  14. J. J. Hull. A database for handwritten text recognition research. IEEE Trans. Pattern Analysis and Machine Intelligence, 16(5):550-554, 1994. 2
  15. I. Jolliffe. Principal Component Analysis. Springer; 2nd edition, 2002. 1
  16. F. D. la Torre Frade and T. Kanade. Discriminative cluster analysis. In ICML, pages 241-248, 2006. 1, 3
  17. S. Ramaswamy, P. Tamayo, R. Rifkin, S. Mukherjee, C.-H. Yeang, M. Angelo, C. Ladd, M. Reich, E. Lat- ulippe, J. Mesirov, T. Poggio, W. Gerald, M. Loda, E. Lander, and T. Golub. Multiclass cancer diagno- sis using tumor gene expression signatures. Proceed- ings of the National Academy of Sciences, 98:15149- 15154, 2001. 2
  18. L. K. Saul and S. T. Roweis. Think globally, fit locally : Unsupervised learning of low dimensional mani- folds. Journal of Machine Learning Research, 4:119- 155, 2003. 1
  19. N. Shental, T. Hertz, D. Weinshall, and M. Pavel. Ad- justment learning and relevant component analysis. In ECCV, pages 776-792, 2002. 1
  20. J. Tenenbaum, V. de Silva, and J. Langford. A global geometric framework for nonlinear dimensionality re- duction. Science, 290(5500):2319-2323, 2000. 1
  21. K. Wagstaff, C. Cardie, S. Rogers, and S. Schroedl. Constrained k-means clustering with background knowledge. In ICML, pages 577-584, 2001. 5
  22. K. Weinberger, J. Blitzer, and L. Saul. Distance metric learning for large margin nearest neighbor classifica- tion. In NIPS, 2005. 1
  23. E. Xing, A. Ng, M. Jordan, and S. Russell. Dis- tance metric learning with application to clustering with side-information. In NIPS, 2002. 1
  24. L. Yang and R. Jin. Distance metric learning: A com- prehensive survey. Technical report, Department of Computer Science and Engineering, Michigan State University, 2006. 1
  25. L. Yang, R. Jin, R. Sukthankar, and Y. Liu. An effi- cient algorithm for local distance metric learning. In AAAI, 2006. 1
  26. D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Schölkopf. Learning with local and global consis- tency. In NIPS, 2004. 6
  27. X. J. Zhu. Semi-supervised learning literature sur- vey. Technical report, Computer Sciences, University of Wisconsin-Madison, 2005. 2, 5
  28. X. J. Zhu, Z. Ghahramani, and J. Lafferty. Semi- supervised learning using gaussian fields and har- monic functions. In ICML, 2003. 6