Academia.eduAcademia.edu

Outline

On ordinal VC-dimension and some notions of complexity

2006, Theoretical Computer Science

https://doi.org/10.1016/J.TCS.2006.07.041

Abstract

We generalize the classical notion of Vapnik-Chernovenkis (VC) dimension to ordinal VC-dimension, in the context of logical learning paradigms. Logical learning paradigms encompass the numerical learning paradigms commonly studied in Inductive Inference. A logical learning paradigm is defined as a set W of structures over some vocabulary, and a set D of first-order formulas that represent data. The sets of models of in W, where varies over D, generate a natural topology W over W. We show that if D is closed under boolean operators, then the notion of ordinal VC-dimension offers a perfect characterization for the problem of predicting the truth of the members of D in a member of W, with an ordinal bound on the number of mistakes. This shows that the notion of VC-dimension has a natural interpretation in Inductive Inference, when cast into a logical setting. We also study the relationships between predictive complexity, selective complexity-a variation on predictive complexity-and mind change complexity. The assumptions that D is closed under boolean operators and that W is compact often play a crucial role to establish connections between these concepts. We then consider a computable setting with effective versions of the complexity measures, and show that the equivalence between ordinal VC-dimension and predictive complexity fails. More precisely, we prove that the effective ordinal VC-dimension of a paradigm can be defined when all other effective notions of complexity are undefined. On a better note, when W is compact, all effective notions of complexity are defined, though they are not related as in the noncomputable version of the framework.

References (13)

  1. A. Ambainis, R. Freivalds, C. Smith, Inductive inference with procrastination: back to definitions, Fund. Inform. 40 (1999) 1-16.
  2. A. Ambainis, S. Jain, A. Sharma, Ordinal mind change complexity of language identification, Theoret. Comput. Sci. 220 (2) (1999) 323-343.
  3. D. Angluin, Inductive inference of formal languages from positive data, Inform. and Control 45 (1980) 117-135.
  4. S. Ben-David, L. Gurvits, A note on VC-dimension and measure of sets of reals, Combin. Probab. Comput. 9 (2000) 391-405.
  5. S. Ben-David, M. Jacovi, On learning in the limit and non-uniform ( , )-learning, in: Proc. Sixth Conf. Computational Learning Theory, ACM Press, New York, 1993, pp. 209-217.
  6. A. Blumer, A. Ehrenfeucht, D. Haussler, M. Warmuth, Learnability and the Vapnik-Chervonenkis dimension, J. ACM 36 (4) (1989) 929-965.
  7. R. Freivalds, C. Smith, On the role of procrastination for machine learning, Inform. and Comput. 107 (2) (1993) 237-271.
  8. E. Gold, Language identification in the limit, Inform. and Control. 10 (1967).
  9. E. Martin, A. Sharma, F. Stephan, A general theory of deduction, induction, and learning, in: K. Jantke, A. Shinohara (Eds.), Proc. Fourth Internat. Conf. on Discovery Science, Lecture Notes in Artificial Intelligence, Vol. 2226, Springer, Berlin, 2001, pp. 228-242.
  10. E. Martin, A. Sharma, F. Stephan, Logic, learning, and topology in a common framework, in: N. Cesa-Bianchi, M. Numao, R. Reischuk (Eds.), Proc. 13th Internat. Conf. on Algebraic Learning Theory, Lecture Notes in Artificial Intelligence, Vol. 2533, Springer, Berlin, 2002, pp. 248-262.
  11. A. Sharma, A note on batch and incremental learnability, J. Comput. System Sci. 56 (1998) 272-276.
  12. L. Valiant, A Theory of the learnable, Commun. ACM 27 (11) (1984) 1134-1142.
  13. V. Vapnik, A. Chervonenkis, On the uniform convergence of relative frequencies of events to their probabilities, Theory Probab. Appl. 16 (2) (1971) 264-280.