Academia.eduAcademia.edu

Outline

Efficient Multiplicative updates for Support Vector Machines

2009, Proceedings of the 2009 SIAM International Conference on Data Mining

https://doi.org/10.1137/1.9781611972795.104

Abstract

The dual formulation of the support vector machine (SVM) objective function is an instance of a nonnegative quadratic programming problem. We reformulate the SVM objective function as a matrix factorization problem which establishes a connection with the regularized nonnegative matrix factorization (NMF) problem. This allows us to derive a novel multiplicative algorithm for solving hard and soft margin SVM. The algorithm follows as a natural extension of the updates for NMF and semi-NMF. No additional parameter setting, such as choosing learning rate, is required. Exploiting the connection between SVM and NMF formulation, we show how NMF algorithms can be applied to the SVM problem. Multiplicative updates that we derive for SVM problem also represent novel updates for semi-NMF. Further this unified view yields algorithmic insights in both directions: we demonstrate that the Kernel Adatron algorithm for solving SVMs can be adapted to NMF problems. Experiments demonstrate rapid convergence to good classifiers. We analyze the rates of asymptotic convergence of the updates and establish tight bounds. We test them on several datasets using various kernels and report equivalent classification performance to that of a standard SVM.

References (19)

  1. Figure 6: Percentage of the misclassification versus the training iteration number (log scale) on the USPS handwritten digits dataset for the M 3 and the MUSIK algorithms (down is better).
  2. Inderjit Dhillon and Suvrit Sra. Generalized nonnega- tive matrix approximations with Bregman divergences. In Y. Weiss, B. Schlkopf, and J. Platt, editors, Ad- vances in Neural Information Processing Systems 18, pages 283-290. MIT Press, Cambridge, MA, 2006.
  3. Chris Ding, Tao Li, and Michael I. Jordan. Convex and semi-nonnegative matrix factorizations. LBNL Tech Report 60428, 2006.
  4. J. Eggert and E. Körner. Sparse coding and NMF. Neural Networks, 2004. Proceedings. 2004 IEEE Inter- national Joint Conference on, 4:2529-2533, 25-29 July 2004.
  5. V. Franc, V. Hlavac, and M. Navara. Sequential coordinate-wise algorithm for the non-negative least squares problem. In Computer Analysis of Images and Patterns, page 407, 2005.
  6. Thilo-Thomas Frieß, Nello Cristianini, and Colin Campbell. The Kernel-Adatron algorithm: a fast and simple learning procedure for support vector machines. In Proc. 15th International Conf. on Machine Learn- ing, pages 188-196. Morgan Kaufmann, San Francisco, CA, 1998.
  7. Vojislav Kecman, Michael Vogt, and Te Ming Huang. On the equality of kernel adatron and sequential mini- mal optimization in classification and regression tasks and alike algorithms for kernel machines. In ESANN, pages 215-222, 2003.
  8. S. S. Keerthi, S. K. Shevade, C. Bhattacharyya, and K. R. K. Murthy. A fast iterative nearest point algorithm for support vector machine classifier design. Neural Networks, IEEE Transactions on, 11(1):124- 136, Jan 2000.
  9. Daniel D. Lee and Sebastian H. Seung. Algorithms for non-negative matrix factorization. In NIPS, pages 556-562, 2000.
  10. Chih-Jen Lin. Projected gradient methods for nonneg- ative matrix factorization. Neural Comp., 19(10):2756- 2779, October 2007.
  11. M. Mørup and L. H. Clemmensen. Multiplicative updates for the LASSO,. Machine Learning for Signal Processing, 2007 IEEE Workshop on, pages 33-38, 2007.
  12. C. L. Blake D. J. Newman and C. J. Merz. UCI repository of machine learning databases, 1998.
  13. E. Osuna, R. Freund, and F. Girosi. An improved training algorithm for support vector machines. Neural Networks for Signal Processing [1997] VII. Proceedings of the 1997 IEEE Workshop, pages 276-285, Sep 1997.
  14. J. Platt. Sequential minimal optimization: A fast algorithm for training support vector machines, 1998.
  15. Ruslan Salakhutdinov and Sam Roweis. Adaptive over- relaxed bound optimization methods. In Proceedings of the International Conference on Machine Learning, volume 20, pages 664-671, 2003.
  16. Bernhard Schölkopf and Alexander J. Smola. Learn- ing with Kernels: Support Vector Machines, Regular- ization, Optimization, and Beyond (Adaptive Compu- tation and Machine Learning). The MIT Press, 2001.
  17. Fei Sha, Lawrence K. Saul, and Daniel D. Lee. Mul- tiplicative updates for large margin classifiers. In Proceedings of the Sixteenth Annual Conference on Computational Learning Theory (COLT), Washington D.C., USA, 2003.
  18. Fei Sha, Lawrence K. Saul, and Daniel D. Lee. Multi- plicative updates for nonnegative quadratic program- ming in support vector machines. In Sebastian Thrun Suzanna Becker and Klaus Obermayer, editors, Ad- vances in Neural Information Processing Systems 15, Cambridge, MA, 2003. MIT Press.
  19. Rafal Zdunek and Andrzej Cichocki. Fast nonnegative matrix factorization algorithms using projected gra- dient approaches for large-scale problems. Computa- tional Intelligence and Neuroscience, page 13, 2008.