On ordinal VC-dimension and some notions of complexity
2006, Theoretical Computer Science
https://doi.org/10.1016/J.TCS.2006.07.041Abstract
We generalize the classical notion of Vapnik-Chernovenkis (VC) dimension to ordinal VC-dimension, in the context of logical learning paradigms. Logical learning paradigms encompass the numerical learning paradigms commonly studied in Inductive Inference. A logical learning paradigm is defined as a set W of structures over some vocabulary, and a set D of first-order formulas that represent data. The sets of models of in W, where varies over D, generate a natural topology W over W. We show that if D is closed under boolean operators, then the notion of ordinal VC-dimension offers a perfect characterization for the problem of predicting the truth of the members of D in a member of W, with an ordinal bound on the number of mistakes. This shows that the notion of VC-dimension has a natural interpretation in Inductive Inference, when cast into a logical setting. We also study the relationships between predictive complexity, selective complexity-a variation on predictive complexity-and mind change complexity. The assumptions that D is closed under boolean operators and that W is compact often play a crucial role to establish connections between these concepts. We then consider a computable setting with effective versions of the complexity measures, and show that the equivalence between ordinal VC-dimension and predictive complexity fails. More precisely, we prove that the effective ordinal VC-dimension of a paradigm can be defined when all other effective notions of complexity are undefined. On a better note, when W is compact, all effective notions of complexity are defined, though they are not related as in the noncomputable version of the framework.
References (13)
- A. Ambainis, R. Freivalds, C. Smith, Inductive inference with procrastination: back to definitions, Fund. Inform. 40 (1999) 1-16.
- A. Ambainis, S. Jain, A. Sharma, Ordinal mind change complexity of language identification, Theoret. Comput. Sci. 220 (2) (1999) 323-343.
- D. Angluin, Inductive inference of formal languages from positive data, Inform. and Control 45 (1980) 117-135.
- S. Ben-David, L. Gurvits, A note on VC-dimension and measure of sets of reals, Combin. Probab. Comput. 9 (2000) 391-405.
- S. Ben-David, M. Jacovi, On learning in the limit and non-uniform ( , )-learning, in: Proc. Sixth Conf. Computational Learning Theory, ACM Press, New York, 1993, pp. 209-217.
- A. Blumer, A. Ehrenfeucht, D. Haussler, M. Warmuth, Learnability and the Vapnik-Chervonenkis dimension, J. ACM 36 (4) (1989) 929-965.
- R. Freivalds, C. Smith, On the role of procrastination for machine learning, Inform. and Comput. 107 (2) (1993) 237-271.
- E. Gold, Language identification in the limit, Inform. and Control. 10 (1967).
- E. Martin, A. Sharma, F. Stephan, A general theory of deduction, induction, and learning, in: K. Jantke, A. Shinohara (Eds.), Proc. Fourth Internat. Conf. on Discovery Science, Lecture Notes in Artificial Intelligence, Vol. 2226, Springer, Berlin, 2001, pp. 228-242.
- E. Martin, A. Sharma, F. Stephan, Logic, learning, and topology in a common framework, in: N. Cesa-Bianchi, M. Numao, R. Reischuk (Eds.), Proc. 13th Internat. Conf. on Algebraic Learning Theory, Lecture Notes in Artificial Intelligence, Vol. 2533, Springer, Berlin, 2002, pp. 248-262.
- A. Sharma, A note on batch and incremental learnability, J. Comput. System Sci. 56 (1998) 272-276.
- L. Valiant, A Theory of the learnable, Commun. ACM 27 (11) (1984) 1134-1142.
- V. Vapnik, A. Chervonenkis, On the uniform convergence of relative frequencies of events to their probabilities, Theory Probab. Appl. 16 (2) (1971) 264-280.