Academia.eduAcademia.edu

Outline

A reinforcement agent for threshold fusion

2008, Applied Soft Computing

https://doi.org/10.1016/J.ASOC.2006.12.003

Abstract

Finding an optimal threshold in order to segment digital images is a difficult task in image processing. Numerous approaches to image thresholding already exist in the literature. In this work, a reinforced threshold fusion for image binarization is introduced which aggregates existing thresholding techniques. The reinforcement agent learns the optimal weights for different thresholds and segments the image globally. A fuzzy reward function is employed to measure object similarities between the binarized image and the original gray-level image, and provide feedback to the agent. The experiments show that promising improvement can be obtained. Three well-established thresholding techniques are combined by the reinforcement agent and the results are compared using error measurements based on ground-truth images.

References (35)

  1. A. Ayesh, Emotionally motivated reinforcement learning based controller, in: Proceedings of the IEEE SMC, The Hague, The Netherlands, 2004.
  2. H.R. Berenji, Fuzzy Q-learning: a new approach for fuzzy dynamic programming problems, in: Proceedings of the Third IEEE International Conference on Fuzzy Systems, Orlando, FL, June 1994.
  3. D. Bhandari, N.R. Pal, D.D. Majumder, Fuzzy divergence, probability measure of fuzzy events and image thresholding, Pattern Recogn. Lett. 13 (1992) 857-867.
  4. K.R. Castleman, Digital Image Processing, Prentice Hall, Upper Saddle River, New Jersey, 1998, 07458.
  5. S. Gadanho, Reinforcement learning in autonomous robots: an empirical investigation of the role of emotions, Ph.D. Thesis, University of Edin- burgh, Edinburgh, 1999.
  6. P.Y. Glorennec, Fuzzy Q-learning and dynamical fuzzy Q-learning, in: Proceedings of the Third IEEE International Conference on Fuzzy Sys- tems, IEEE Press, Piscataway, NJ, 1994, pp. 474-479.
  7. P.Y. Glorennec, L. Jouffe, Fuzzy Q-learning, in: Proceedings of the Sixth International Conference on Fuzzy Systems, Barcelona, Spain, (1997), pp. 659-662.
  8. R.C. Gonzalez, R.E. Woods, Digital Image Processing, 2nd ed., Prentice Hall, New Jersey, 2002, 07458.
  9. W.D. Gregg, Analog and Digital Communication, John Wiley and Sons, Inc., USA, 1977.
  10. R.W. Hamming, Error detecting and error correcting codes, Bell System Tech. J. 29 (1950) 933-968.
  11. L.K. Huang, M.J.J. Wang, Image thresholding by minimizing the mea- sures of fuzziness, Pattern Recogn. 28 (1995) 41-51.
  12. L. Jouffe, Fuzzy inference system learning bay reinforcement methods, IEEE Trans. Syst. Man Cybern. 28 (1998) 338-355.
  13. L.P. Kaelbling, M.L. Littman, A.W. Moore, Reinforcement learning: a survey, J. Artif. Intell. Res. 4 (1996) 237-285.
  14. J.N. Kapur, P.K. Sahoo, A.K.C. Wong, A new method for gray-level picture thresholding using the entropy of the histogram, Comput. Vis. Graph. Image Process. 29 (1985) 273-285.
  15. J. Kittler, J. Illingworth, Minimum error thresholding, Pattern Recogn. 19 (1) (1986) 41-47.
  16. S.U. Lee, S.Y. Chung, R.H. Park, A Comparative performance study of several global thresholding techniques for segmentation, Comput. Vis. Graph. Image Process. 52 (1990) 171-190.
  17. N. Otsu, A threshold selection method from gray level histogram, IEEE Trans. Syst. Man Cybern. SMC-9 (1979) 62-66.
  18. S.J. Russell, P. Norvig, Artificial Intelligence: A Modern Approach, Prentice Hall, Englewood Cliffs, N.J., 1995.
  19. P.K. Sahoo, S. Soltani, A.K.C. Wong, Y.C. Chen, A survey of thresholding technique, Comput. Vis. Graph. Image Process. 41 (1988) 233-260.
  20. B. Sankur, M. Sezgin, Survey over image thresholding techniques and quantitative performance evaluation, J. Electron. Imaging 13 (1) (2004) 146-165.
  21. M. Shokri, H.R. Tizhoosh, Using reinforcement learning for image thresholding, Can. Conf. Electr. Comput. Eng. 1 (2003) 1231-1234.
  22. M. Shokri, H.R. Tizhoosh, Q(l)-based image thresholding, Can. Conf. Comput. Robot Vis. (2004) 504-508.
  23. R.S. Sutton, A.G. Barto, Reinforcement Learning: An Introduction, MIT Press, Cambridge, Mass, 1998.
  24. W. Tsai, Moment-preserving thresholding: a new approach, Comput. Vis. Graph. Image Process. 29 (1985) 377-393.
  25. M.A. Walker, An application of reinforcement learning to dialogue strategy selection in a spoken dialogue system for email, J. Artif. Intell. Res. 12 (2000) 387-416.
  26. C.J.C.H. Watkins, Learning from Delayed Rewards, Cambridge Uni- versity, Cambridge, 1989.
  27. C.J.H. Watkins, P. Dayan, Technical note, Q-learning, Mach. Learn. 8 (1992) 279-292.
  28. H. Yan, Unified formulation of a class of image thresholding techniques, Pattern Recogn. 29 (12) (1996) 2025-2032.
  29. P. Yin, Maximum entropy-based optimal threshold selection using deter- ministic reinforcement learning with controlled randomization, Signal Process. 82 (2002) 993-1006.
  30. J. Yoshimoto, Sh. Ishii, M. Sato, Application of reinforcement learning to balancing of acrobat, IEEE Int. Conf. Syst. Man Cybern. (2003) 516-521.
  31. W. Zhang, T.G. Dietterich, Value function approximations and job-shop scheduling, Workshop on Value Function Approximation in Reinforce- ment Learning at ICML-95, 1995, submitted.
  32. S. Ali, K.A. Smith, On learning algorithm selection for classification, Appl. Soft Comput. J. 6 (2) (2006) 119-138.
  33. I. Ko ´kai, A. Lo ¨rincz, Fast adapting value estimation-based hybrid archi- tecture for searching the world-wide web, Appl. Soft Comput. J. 2 (1) (2002) 11-23.
  34. M. Shokri, H.R. Tizhoosh, M. Kamel, The outline of reinforcement learning technique for e-learning applications, in: S. Pierre (Ed.), E- Learning Networked Environments and Architectures: A Knowledge Processing Perspective, Springer Book Series, 2007.
  35. B. Bhanu, Y. Lin, Object detection in multi-modal images using genetic programming, Appl. Soft Comput. J. 4 (2) (2004) 175-201.