Ensemble Sparse Models for Image Analysis
Abstract
Sparse representations with learned dictionaries have been successful in several image analysis applications. In this paper, we propose and analyze the framework of ensemble sparse models, and demonstrate their utility in image restoration and unsupervised clustering. The proposed ensemble model approximates the data as a linear combination of approximations from multiple weak sparse models. Theoretical analysis of the ensemble model reveals that even in the worst-case, the ensemble can perform better than any of its constituent individual models. The dictionaries corresponding to the individual sparse models are obtained using either random example selection or boosted approaches. Boosted approaches learn one dictionary per round such that the dictionary learned in a particular round is optimized for the training examples having high reconstruction error in the previous round. Results with compressed recovery show that the ensemble representations lead to a better performance compared to using a single dictionary obtained with the conventional alternating minimization approach. The proposed ensemble models are also used for single image superresolution, and we show that they perform comparably to the recent approaches. In unsupervised clustering, experiments show that the proposed model performs better than baseline approaches in several standard datasets.
References (45)
- D. Field, "What is the goal of sensory coding?" Neural computation, vol. 6, no. 4, pp. 559-601, 1994.
- M. Elad, Sparse and redundant representations: from theory to appli- cations in signal and image processing. Springer, 2010.
- J. Thiagarajan, K. Ramamurthy, and A. Spanias, "Multilevel dictionary learning for sparse representation of images," in IEEE DSPE Workshop, 2011, pp. 271-276.
- J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma, "Robust face recognition via sparse representation," IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 31, no. 2, pp. 210-227, 2009.
- I. Ramirez, P. Sprechmann, and G. Sapiro, "Classification and clustering via dictionary learning with structured incoherence and shared features," in IEEE CVPR, 2010, pp. 3501-3508.
- J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, "Supervised dictionary learning," Advances in neural information processing systems, 2008.
- D. L. Donoho and M. Elad, "Optimally sparse representation in general (nonorthogonal) dictionaries via l minimization." Proceedings of the National Academy of Sciences of the United States of America, vol. 100, no. 5, pp. 2197-202, Mar. 2003.
- S. Mallat and Z. Zhang, "Matching pursuits with time-frequency dictio- naries," IEEE Transactions on Signal Processing, vol. 41, no. 12, pp. 3397-3415, 1993.
- J. A. Tropp, "Greed is good: Algorithmic results for sparse approxima- tion," IEEE Transactions on Information Theory, vol. 50, no. 10, pp. 2231-2242, October 2004.
- S. Cotter, R. Adler, R. Rao, and K. Kreutz-Delgado, "Forward sequential algorithms for best basis selection," in Vision, Image and Signal Pro- cessing, IEE Proceedings-, vol. 146, no. 5. IET, 1999, pp. 235-244.
- S. S. Chen, D. L. Donoho, and M. A. Saunders, "Atomic decomposition by basis pursuit," SIAM Review, vol. 43, no. 1, pp. 129-159, 2001.
- I. F. Gorodnitsky and B. D. Rao, "Sparse signal reconstruction from lim- ited data using FOCUSS: A re-weighted norm minimization algorithm," IEEE Transactions on Signal Processing, vol. 45, no. 3, pp. 600-616, March 1997.
- M. Elad, "Why simple shrinkage is still relevant for redundant repre- sentations?" IEEE Transactions on Information Theory, vol. 52, no. 12, pp. 5559 -5569, December 2006.
- M. Elad et.al., "A wide-angle view at iterated shrinkage algorithms," in SPIE (Wavelet XII) 2007, 2007.
- J. Mairal, F. Bach, J. Ponce, and G. Sapiro, "Online dictionary learning for sparse coding," in Proc. ICML, 2009, pp. 689-696.
- B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, "Least angle regression," The Annals of statistics, vol. 32, no. 2, pp. 407-499, 2004.
- H. Lee, A. Battle, R. Raina, and A. Ng, "Efficient sparse coding algorithms," Advances in neural information processing systems, vol. 19, p. 801, 2007.
- M. Aharon, M. Elad, and A. Bruckstein, "The K-SVD: an algorithm for designing of overcomplete dictionaries for sparse representation," IEEE Trans. Signal Processing, vol. 54, no. 11, pp. 4311-4322, 2006.
- M. Aharon and M. Elad, "Image denoising via sparse and redundant representations over learned dictionaries," IEEE Transactions on Image Processing, vol. 15, no. 12, pp. 3736-3745, 2006.
- J. Yang, J. Wright, T. Huang, and Y. Ma, "Image super-resolution as sparse representation of raw image patches," in Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on. IEEE, 2008, pp. 1-8.
- K. Huang and S. Aviyente, "Sparse representation for signal classifica- tion," in Proc. of Advances in Neural Information Processing Systems 19. MIT Press, 2006.
- J. J. Thiagarajan, K. N. Ramamurthy, and A. Spanias, "Sparse repre- sentations for pattern classification using learned dictionaries," Proc. of Twenty-eighth SGAI International Conference on Artificial Intelligence, 2008.
- J. J. Thiagarajan, K. N. Ramamurthy, P. Knee, and A. Spanias, "Sparse representations for automatic target classification in SAR images," in Proc. of ISCCSP, 2010.
- J. Yang et.al., "Linear spatial pyramid matching using sparse coding for image classification," in IEEE CVPR, 2009.
- G. Yu, G. Sapiro, and S. Mallat, "Image modeling and enhancement via structured sparse model selection," in Proc. of IEEE ICIP, Sep. 2010, pp. 1641 -1644.
- Q. Zhang and B. Li, "Discriminative K-SVD for dictionary learning in face recognition," in IEEE CVPR, 2010.
- J. J. Thiagarajan and A. Spanias, "Learning dictionaries for local sparse coding in image classification," in Proc. of Asilomar SSC, 2011.
- J. J. Thiagarajan, K. N. Ramamurthy, P. Sattigeri, and A. Spanias, "Su- pervised local sparse coding of sub-image features for image retrieval," in IEEE ICIP, 2012.
- B. Cheng, J. Yang, S. Yan, Y. Fu, and T. S. Huang, "Learning with l1-graph for image analysis." IEEE transactions on image processing : a publication of the IEEE Signal Processing Society, vol. 19, no. 4, pp. 858-66, Apr. 2010. [Online]. Available: http://www.ncbi.nlm.nih.gov/pubmed/20031500
- Y. Freund, R. Schapire, and N. Abe, "A short introduction to boosting," Journal-Japanese Society For Artificial Intelligence, vol. 14, no. 771- 780, p. 1612, 1999.
- R. Polikar, "Ensemble based systems in decision making," Circuits and Systems Magazine, IEEE, vol. 6, no. 3, pp. 21-45, 2006.
- W. Zhang, A. Surve, X. Fern, and T. Dietterich, "Learning non-redundant codebooks for classifying complex objects," in Proc. ICML, 2009, pp. 1241-1248.
- J. Wang, Y. Li, Y. Zhang, H. Xie, and C. Wang, "Boosted learning of visual word weighting factors for bag-of-features based medical image retrieval," in International Conference on Image and Graphics, 2011, pp. 1035-1040.
- M. Elad and I. Yavneh, "A plurality of sparse representations is better than the sparsest one alone," Information Theory, IEEE Transactions on, vol. 55, no. 10, pp. 4701-4714, 2009.
- N. Duffy and D. Helmbold, "Boosting Methods for Regression," pp. 153-200, 2002.
- J. Yang, J. Wright, T. S. Huang, and Y. Ma, "Image super-resolution via sparse representation," Image Processing, IEEE Transactions on, vol. 19, no. 11, pp. 2861-2873, 2010.
- T. Dietterich, "Ensemble methods in machine learning," Multiple clas- sifier systems, pp. 1-15, 2000.
- "ScSR -matlab codes for image super-resolution," Available at http://www.ifp.illinois.edu/∼jyang29/resources.html.
- "Berkeley segmentation dataset," Available at http://www.eecs.berkeley.edu/Research/Projects/CS/ vi- sion/grouping/segbench/.
- B. Bahmani, B. Moseley, A. Vattani, R. Kumar, and S. Vassilvitskii, "Scalable k-means++," Proceedings of the VLDB Endowment, vol. 5, no. 7, pp. 622-633, 2012.
- D. Arthur and S. Vassilvitskii, "K-means++: The advantages of careful seeding," in Proc. ACM-SIAM symposium on Discrete algorithms, 2007, pp. 1027-1035.
- D. Donoho, "Compressed sensing," IEEE Trans. Information Theory, vol. 52, no. 4, pp. 1289-1306, 2006.
- T. G. Dietterich, "An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and ran- domization," Machine learning, vol. 40, no. 2, pp. 139-157, 2000.
- E. Elhamifar and R. Vidal, "Sparse subspace clustering: Algorithm, theory, and applications," arXiv preprint arXiv:1203.1005, 2012.
- A. Y. Ng, M. I. Jordan, Y. Weiss et al., "On spectral clustering: Analysis and an algorithm," Advances in neural information processing systems, vol. 2, pp. 849-856, 2002.