Dynamics of the evolution of learning algorithms by selection
2003, Physical Review E
https://doi.org/10.1103/PHYSREVE.67.041912Abstract
We study the evolution of artificial learning systems by means of selection. Genetic programming is used to generate populations of programs that implement algorithms used by neural network classifiers to learn a rule in a supervised learning scenario. In contrast to concentrating on final results, which would be the natural aim while designing good learning algorithms, we study the evolution process. Phenotypic and genotypic entropies, which describe the distribution of fitness and of symbols, respectively, are used to monitor the dynamics. We identify significant functional structures responsible for the improvements in the learning process. In particular, some combinations of variables and operators are useful in assessing performance in rule extraction and can thus implement annealing of the learning schedule. We also find combinations that can signal surprise, measured on a single example, by the difference between predicted and correct classification. When such favorable structures appear, they are disseminated on very short time scales throughout the population. Due to such abruptness they can be thought of as dynamical transitions. But foremost, we find a strict temporal order of such discoveries. Structures that measure performance are never useful before those for measuring surprise. Invasions of the population by such structures in the reverse order were never observed. Asymptotically, the generalization ability approaches Bayesian results.
References (17)
- J. R. Koza, Genetic Programming: On the Programming of Computers by Means of Natural Selection ͑MIT Press, Cam- bridge, MA, 1992͒.
- J. H. Holland, Adaptations in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control and Artificial Intelligence ͑University of Michigan Press, Ann Arbor, 1975͒.
- S. Amari, IEEE Trans. Electron. Comput. 16, 299 ͑1967͒.
- A. Engel and C. Van den Broeck, Statistical Mechanics of Learning ͑Cambridge University Press, Cambridge, 2000͒.
- M. Opper and D. Haussler, Phys. Rev. Lett. 66, 2677 ͑1991͒.
- O. Kinouchi and N. Caticha, J. Phys. A 25, 6243 ͑1992͒.
- M. Copelli and N. Caticha, J. Phys. A 28, 1615 ͑1994͒.
- R. Vicente and N. Caticha, J. Phys. A 30, L599 ͑1997͒.
- R. Simonetti and N. Caticha, J. Phys. A 29, 6243 ͑1996͒.
- M. Biehl and H. Schwarze, Europhys. Lett. 20, 733 ͑1992͒.
- O. Kinouchi and N. Caticha, J. Phys. A 26, 6161 ͑1993͒. ͓12͔ C. Van den Broeck and P. Reimann, Phys. Rev. Lett. 76, 2188 ͑1996͒.
- M. Opper, Phys. Rev. Lett. 77, 4671 ͑1996͒.
- M. Opper, in On-line Learning in Neural Networks, edited by D. Saad ͑Cambridge University Press, Cambridge, 1998͒, p. 363.
- S. Sara and O. Winther, in On-line Learning in Neural Net- works ͑Ref. ͓14͔͒, p. 379.
- O. Kinouchi and N. Caticha, Phys. Rev. E 54, R54 ͑1996͒.
- N. Caticha and O. Kinouchi, Philos. Mag. B 77, 1565 ͑1998͒. ͓18͔ This is similar to that introduced in C. Adami, C. Ofria, and T. C. Collier, Proc. Natl. Acad. Sci. U.S.A. 97, 4463 ͑2000͒, but not restricted to fixed size program length. ͓19͔ Advances in Genetic Programming, edited by K. E. Kinnear, Jr. ͑MIT Press, Cambridge, MA, 1994͒.
- J. P. Neirotti and N. Caticha, available at http:// www.fge.if.usp.brϳnestor/