HAL (Le Centre pour la Communication Scientifique Directe), Jun 1, 2015
Nous présentons TermLis un contexte d'information logique construit à partir de ressources termin... more Nous présentons TermLis un contexte d'information logique construit à partir de ressources terminologiques disponibles en xml (FranceTerme), pour une utilisation flexible avec un logiciel de contexte logique (CAMELIS). Une vue en contexte logique permet d'explorer des informations de manière flexible, sans rédaction de requête a priori, et d'obtenir aussi des indications sur la qualité des données. Un tel contexte peut être enrichi par d'autres informations (de natures diverses), mais aussi en le reliant à d'autres applications (par des actions associées selon des arguments fournis par le contexte). Nous montrons comment utiliser TermLis et nous illustrons, à travers cette réalisation concrète sur des données de FranceTerme, les avantages d'une telle approche pour des données terminologiques.
This paper explains new results relating modal propositional logic and rewrite rule systems. More... more This paper explains new results relating modal propositional logic and rewrite rule systems. More precisely, we give complete term rewriting systems for the modal propositional systems known as K, Q, T, and S5. These systems are presented as extensions of Hsiang's system for classical propositional calculus. We have checked local confluence with the rewrite rule system K.B. (cf. the Knuth-Bendix algorithm) developed by the Formel project at INRIA. We prove that these systems are noetherian, and then infer their confluence from Newman's lemma. Therefore each term rewriting system provides a new automated decision procedure and defines a canonical form for the corresponding logic. We also show how to characterize the canonical forms thus obtained. normal modal systems; the system K is in a precise sense the weakest normal modal system. These systems are based on classical propositional calculus (CPC). We briefly recall their definitions as Hilbert systems.
This paper is concerned with learning categorial grammars from positive examples in the model of ... more This paper is concerned with learning categorial grammars from positive examples in the model of Gold. Functor-argument structures (written FA) are usual syntactical decompositions of sentences in sub-components distinguishing the functional parts from the argument parts defined in the case of classical categorial grammars also known as AB-grammars. In the case of nonassociative type-logical grammars, we propose a similar notion that we call generalized functor-argument structures and we show that these structures capture the essence of non-associative Lambek (NL) calculus without product. We show that (i) rigid and k-valued non-associative Lambek (NL without product) grammars are learnable from generalized functor-argument structured sentences. We also define subclasses of k-valued grammars in terms of arity. We first show that (ii) for each k and each bound on arity the class of FA-arity bounded k-valued NL languages of FA structures is finite and (iii) that FA-arity bounded k-valued NL grammars are learnable both from strings and from FA structures as a corollary. Result (i) is obtained from (ii); this learnability result (i) is interesting and surprising when compared to other results: in fact we also show that (iv) this class has infinite elasticity. Moreover, these classes are very close to classes like rigid associative Lambek grammars learned from natural deduction structured sentences (that are different and much richer than FA or generalized FA) or to k-valued non-associative Lambek grammars unlearnable from strings or even from bracketed strings. Thus, the class of k-valued non-associative Lambek grammars learned from generalized functor-argument sentences is at the frontier between learnable and unlearnable classes of languages.
HAL (Le Centre pour la Communication Scientifique Directe), Dec 1, 2021
Requirement documents, including natural language statement sets, form the basis for system desig... more Requirement documents, including natural language statement sets, form the basis for system design. Their validation is a critical task. Formal methods exist to assist the designer in this way, but they appear too complex for practical use. This paper presents a new methodology that aims at reducing this gap. It is based on graphbased meaning representations and compositional transduction operations, so as to extract and formalize requirements in suitable formats. Our targets are automata enabling efficient formal reasoning. This proposal is also intended to be simple, reliable and transparent.
We describe the theoretical principles that underlie the design of a software tool which could be... more We describe the theoretical principles that underlie the design of a software tool which could be used by judges for writing judgements and for making decisions about litigations. The tool is based on Binary Decision Diagrams (BDD), which are graphical representations of truth–valued functions associated to propositional formulas. Given a specific litigation, the tool asks questions to the judge; each question is represented by a propositional atom. Their answers, true or false, allow to evaluate the truth value of the formula which encodes the overall recommendation of the software about the litigation. Our approach combines some sort of ‘theoretical’ or ‘legal’ reasoning dealing with the core of the litigation itself together with some sort of ‘procedural’ reasoning dealing with the protocol that has to be followed by the judge during the trial: some questions or group of questions must necessarily be examined and sometimes in a specific order. That is why we consider extensions o...
We describe the theoretical principles that underlie the design of a software tool which could be... more We describe the theoretical principles that underlie the design of a software tool which could be used by judges for writing judgements and for making decisions about litigations. The tool is based on Binary Decision Diagrams (BDD), which are graphical representations of truth-valued functions associated to propositional formulas. Given a specific litigation, the tool asks questions to the judge; each question is represented by a propositional atom. Their answers, true or false, allow to evaluate the truth value of the formula which encodes the overall recommendation of the software about the litigation. Our approach combines some sort of 'theoretical' or 'legal' reasoning dealing with the core of the litigation itself together with some sort of 'procedural' reasoning dealing with the protocol that has to be followed by the judge during the trial: some questions or group of questions must necessarily be examined and sometimes in a specific order. That is why ...
HAL (Le Centre pour la Communication Scientifique Directe), Dec 1, 2021
Requirement documents, including natural language statement sets, form the basis for system desig... more Requirement documents, including natural language statement sets, form the basis for system design. Their validation is a critical task. Formal methods exist to assist the designer in this way, but they appear too complex for practical use. This paper presents a new methodology that aims at reducing this gap. It is based on graphbased meaning representations and compositional transduction operations, so as to extract and formalize requirements in suitable formats. Our targets are automata enabling efficient formal reasoning. This proposal is also intended to be simple, reliable and transparent.
Nous presentons dans cet article les methodes employees et les resul-tats obtenus en reponse au D... more Nous presentons dans cet article les methodes employees et les resul-tats obtenus en reponse au Defi EGC 2016. Notre approche repose d'une part sur des chaines automatiques de traitements linguistiques en francais et en an-glais utilisant le plus possible des ressources et outils publics et d'autre part sur un environnement d'exploration des donnees base sur les systemes d'informa-tion logiques ; ces systemes exploitent une generalisation des treillis de concepts formels appliquee aux donnees attribut-valeur ou au web semantique.
We study some learnability problems in the family of Categorial Dependency Grammars (CDG), a clas... more We study some learnability problems in the family of Categorial Dependency Grammars (CDG), a class of categorial grammars defining dependency structures. CDG is a formal system, where types are attached to words, combining the classical categorial grammars' elimination rules with valency pairing rules defining non-projective (discontinuous) dependencies; very importantly, the elimination rules are naturally extended to the so called "iterated dependencies" expressed by a specific type constructor and related elimination rules. This paper first reviews key points on negative results: even the rigid (one type per word) CDG cannot be learned neither from function/argument structures, nor even from dependency structures themselves. Such negative results prove the impossibility to define a learning algorithm for these grammar classes. Nevertheless, we show that the CDG satisfying reasonable and linguistically valid conditions on the iterated dependencies are incrementally learnable in the limit from dependency structures. We provide algorithms and also discuss these aspects for recent variants of the formalism that allow the inference of CDG from linguistic treebanks.
Le projet LangNum-br-fr 1 concerne la paire de langues francais-breton et le numerique, dans un c... more Le projet LangNum-br-fr 1 concerne la paire de langues francais-breton et le numerique, dans un cadre pluridisciplinaire. Il regroupe des informaticiens specialistes en traitement automatique des langues (Univ. Rennes & IRISA, LIG Grenoble, Univ Tours), des linguistes specialistes des langues celtiques (CRBC) et des specialistes des usages des TIC 2 (LOUSTIC Rennes2), en vue de valoriser des ressources pour le breton et avec une preoccupation pedagogique. Une analyse des besoins des apprenants en est la premiere etape avant la definition de traitements logiciels pour des meilleurs et nouveaux usages.
The concept of pregroup was introduced by Lambek for natural language analysis, with a close link... more The concept of pregroup was introduced by Lambek for natural language analysis, with a close link to non-commutative linear logic. We reformulate the pregroup calculus so as to extend it by composition with other logics and calculi. The cut elimination property and the decidability property of the sequent calculus proposed in the article are shown. Properties of composed calculi are also discussed.
Pregroup grammars are a context-free grammar formalism which may be used to describe the syntax o... more Pregroup grammars are a context-free grammar formalism which may be used to describe the syntax of natural languages. However, this formalism is not able to naturally define the types of optional or iterated arguments like optional verb complements or its iterated optional circumstantials. In this paper are introduced and formalized two constructions that make up for this inadequacy without loss of efficiency.
We propose a novel subclass in the family of Categorial Dependency Grammars (CDG), based on a syn... more We propose a novel subclass in the family of Categorial Dependency Grammars (CDG), based on a syntactic criterion on categorial types associated to words in the lexicon and study its learnability. This proposal relies on a linguistic principle and relates to a former non-constructive condition on iterated dependencies. We show that the projective CDG in this subclass are incrementally learnable in the limit from dependency structures. In contrast to previous proposals, our criterion is both syntactic and does not impose a (rigidity) bound on the number of categorial types associated to a word.
Pregroup grammars are a context-free grammar formalism which may be used to describe the syntax o... more Pregroup grammars are a context-free grammar formalism which may be used to describe the syntax of natural languages. However, this formalism is not able to naturally define types corresponding to optional and iterated arguments such as optional complements of verbs or verbs’ adverbial modifiers. This paper introduces two constructions that make up for this deficiency.
Les grammaires de prégroupes (PG) ( ?) ont été introduites comme une simplification du calcul de ... more Les grammaires de prégroupes (PG) ( ?) ont été introduites comme une simplification du calcul de Lambek (?). Elles ont été utilisées pour modéliser des fragments de la syntaxe de plusieurs langages naturels : anglais ( ?), italien (?), français ( ?), allemand (?; ?), japonais ( ?), perse ( ?), etc. Elles appartiennent à la classes des grammaires catégorielles et sont fortement lexicalisées : les grammaires catégorielles ont des liens privilégiés avec l’interprétation sémantique tandis que l’aspect lexicalisé apporte des avantages pour la construction des grammaires et pour l’analyse syntaxique. Un autre interêt des PG réside dans leur possibilité de définir un ordre sur les types primitifs, ce qui aide à la conception des grammaires en introduisant des types compacts et moins nombreux (comparées à d’autres types de grammaires catégorielles par exemple). Ce point permet aussi de combiner les systèmes de manière hiérarchisée ( ?). De plus, contrairement à des variantes de grammaires c...
Recently, learning algorithms in Gold’s model have been proposed for some particular classes of c... more Recently, learning algorithms in Gold’s model have been proposed for some particular classes of classical categorial grammars [Kan98]. We are interested here in learning Lambek categorial grammars. In general grammatical inference uses unification and substitution. In the context of Lambek categorial grammars it seems appropriate to incorporate an operation on types based both on deduction (Lambek derivation) and on substitution instead of standard substitution and standard unification. After an introduction (in connection with learning), this paper will recall conjoinability results [Lam58, Pen93] using groups ; we then consider a characterization of conjoinability using quasi-groups for the nonassociative version of Lambek calculus. We then relate these characterizations to the modified unification investigated in [For01] for the associative Lambek calculus.
Akenou-Breizh, a platform project to develop computational and linguistic resources and tools for... more Akenou-Breizh, a platform project to develop computational and linguistic resources and tools for breton We present a new project, Akenou-Breizh, that aims to (1) put in place a platform allowing to study the influences of an heritage language, such as Breton, on a usage language, such as French, and (2) to make available, to all interested persons, tools well integrated in the “semantic and multilingual web” and proposing proactive access to various kinds of knowledge concerning Breton, as well as direct visualisation of infrasentential correspondences in aligned bilingual presentations. We plan not only to use the numerous freely available resources, in particular those of OPLB and of the APERTIUM project, but also to create new ones, such as good quality bilingual aligned corpora, thereby using the “collaborative web”, and to build on the dedicated lingwarium.org web site linguistic modules improving on or extending those that exist, for example a morphological analyzer-generator...
Uploads
Papers by Annie Foret