BACK TO INDEX

All publications sorted by year
2022
  1. Denis Béchet and Annie Foret. Incremental learning of iterated dependencies. Journal of Machine Learning, 2022. [doi:10.1007/s10994-021-05947-2] Keyword(s): Grammatical inference - Categorial grammar - Dependency grammar - Incremental learning - Iterated types - Treebanks.
    Abstract:
    We study some learnability problems in the family of Categorial Dependency Grammars (CDG), a class of categorial grammars defining dependency structures. CDG is a formal system, where types are attached to words, combining the classical categorial grammars' elimination rules with valency pairing rules defining non-projective (discontinuous) dependencies; very importantly, the elimination rules are naturally extended to the so called "iterated dependencies" expressed by a specific type constructor and related elimination rules. This paper first reviews key points on negative results: even the rigid (one type per word) CDG cannot be learned neither from function/argument structures, nor even from dependency structures themselves. Such negative results prove the impossibility to define a learning algorithm for these grammar classes. Nevertheless, we show that the CDG satisfying reasonable and linguistically valid conditions on the iterated dependencies are incrementally learnable in the limit from dependency structures. We provide algorithms and also discuss these aspects for recent variants of the formalism that allow the inference of CDG from linguistic treebanks.
    [bibtex-key = foret:ML2021] [bibtex-entry]


  2. Mireille Ducassé. Kartu-Verbs : un système d'informations logiques de formes verbales fléchies pour contourner les problèmes de lemmatisation des verbes géorgiens. Revue des Nouvelles Technologies de l'Information, Extraction et Gestion des Connaissances, RNTI-E-38:421-428, 2022. Note: Démonstration. [WWW]
    Abstract:
    La langue géorgienne possède un système verbal complexe, à la fois agglutinant et flexionnel, avec de nombreusesirrégularités. Les formes fléchies d'un verbe peuvent être très différentes les unes des autres. Il faut une bonne connaissance de la grammaire géorgienne pour remonter à l'infinitif (le lemme d'accès des dictionnaires le plus fréquent). L'accès aux dictionnaires pour les débutants est, de ce fait, très difficile. De plus, il n'y a pas de consensus parmi les lexicographes du Géorgien sur les lemmes qui représentent un verbe dans les dictionnaires,ce qui complexifie encore davantage les accès.Nous proposons Kartu-Verbs, une base de formes fléchies de verbes géorgiens accessible par un système d'informations logiques. Cette démonstration montre comment, à partir de n'importe quelle forme fléchie, on peut trouver le lemme pertinent pour accéder à n'importe quel dictionnaire. Kartu-Verbs peut, ainsi,être utilisé comme une interface aux dictionnaires géorgiens.
    [bibtex-key = ducasse2022] [bibtex-entry]


  3. Anne-Laure Ligozat, Kevin Marquet, Aurélie Bugeau, Julien Lefevre, Pierre Boulet, Sylvain Bouveret, Philippe Marquet, Olivier Ridoux, and Olivier Michel. How to Integrate Environmental Challenges in Computing Curricula?. In Proceedings of the 53rd ACM Technical Symposium on Computer Science Education (SIGCSE'22), Providence, RI, USA., March 2022. ACM. [WWW]
    Abstract:
    This paper advocates for the integration of environmental aspects in computing curricula, with a focus on higher education. We created knowledge-based curriculum specifications in order to help teachers who wish to add knowledge foundation on computing impacts. This document lists topics and references that can be integrated into curricula. We implemented it in several higher education institutions. This paper reports on our experience and feedback. We also discuss recommendations to overcome obstacles that, from our experience, are often faced when modifying computing curricula to integrate environmental challenges.
    [bibtex-key = LMBLBBMRM-SIGCSE22] [bibtex-entry]


2021
  1. Francesco Bariatti. Mining Tractable Sets of Graph Patterns with the Minimum Description Length Principle. PhD thesis, Université de Rennes 1, November 2021. [WWW] [bibtex-key = Bar2021Thesis] [bibtex-entry]


  2. Aurélien Lamercerie. Principe de transduction sémantique pour l'application de théories d'interfaces sur des documents de spécification. PhD thesis, Université de Rennes 1, April 2021. Note: Thèse de doctorat dirigée par Caillaud, Benoit et Foret, Annie. [WWW]
    Abstract:
    The specification of technical systems is a complex and error-prone task. From a methodological point of view, the expected characteristics must be rigorously specified. In practice, specifications group together the desired properties in the form of a list of rules to be checked, called requirements. The challenge of this book is to build an analysis process for application on textual documents, written in a natural language such as English. The targeted implementation is an end-to-end automated processing chain, and integrating faculties of interpretation and reasoning on specification data. Precisely, we propose to study and experiment how to link natural language statements to formal models that can be exploited in a theoretical framework. First, the principle of semantic transduction is advanced to extract and formalize natural language statements. In a second step, the algebraic properties of specification models are studied to define a theory to verify the consistency of requirements.
    [bibtex-key = lamer2021] [bibtex-entry]


  3. Sébastien Ferré. Conceptual Navigation in Large Knowledge Graphs. In Rokia Missaoui, Leonard Kwuida, and Talel Abdessalem, editors, Complex Data Analysis with Formal Concept Analysis. Springer, 2021. Note: To appear. Keyword(s): knowledge graph, formal concept analysis, Graph-FCA, conceptual navigation.
    Abstract:
    A growing part of Big Data is made of knowledge graphs. Major knowledge graphs such as Wikidata, DBpedia or the Google Knowledge Graph count millions of entities and billions of semantic links. A major challenge is to enable their exploration and querying by end-users. The SPARQL query language is powerful but provides no support for exploration by end-users. Question answering is user-friendly but is limited in expressivity and reliability. Navigation in concept lattices supports exploration but is limited in expressivity and scalability. % In this paper, we introduce a new exploration and querying paradigm, Abstract Conceptual Navigation (ACN), that merges querying and navigation in order to reconcile expressivity, usability, and scalability. ACN is founded on Formal Concept Analysis (FCA) by defining the navigation space as a concept lattice. We then instantiate the ACN paradigm to knowledge graphs (Graph-ACN) by relying on Graph-FCA, an extension of FCA to knowledge graphs. We continue by detailing how Graph-ACN can be efficiently implemented on top of SPARQL endpoints, and how its expressivity can be increased in a modular way. Finally, we present a concrete implementation available online, Sparklis, and a few application cases on large knowledge graphs.
    [bibtex-key = Fer2021cda_fca] [bibtex-entry]


  4. Sébastien Ferré. Application of Concepts of Neighbours to Knowledge Graph Completion. Data Science: Methods, Infrastructure, and Applications, 4:1-28, 2021. [doi:10.3233/DS-200030] Keyword(s): knowledge graph, link prediction, concepts of enighbours.
    Abstract:
    The open nature of Knowledge Graphs (KG) often implies that they are incomplete. Knowledge graph completion (aka. link prediction) consists in inferring new relationships between the entities of a KG based on existing relationships. Most existing approaches rely on the learning of latent feature vectors for the encoding of entities and relations. In general however, latent features cannot be easily interpreted. Rule-based approaches offer interpretability but a distinct ruleset must be learned for each relation. In both latent- and rule-based approaches, the training phase has to be run again when the KG is updated. We propose a new approach that does not need a training phase, and that can provide interpretable explanations for each inference. It relies on the computation of Concepts of Nearest Neighbours (C-NN) to identify clusters of similar entities based on common graph patterns. Different rules are then derived from those graph patterns, and combined to predict new relationships. We evaluate our approach on standard benchmarks for link prediction, where it gets competitive performance compared to existing approaches.
    [bibtex-key = Fer2020ds] [bibtex-entry]


  5. Hugo Ayats, Peggy Cellier, and Sébastien Ferré. Extracting Relations in Texts with Concepts of Neighbours. In Agnès Braud, Aleksey Buzmakov, Tom Hanika, and Florence Le Ber, editors, Formal Concept Analysis, LNCS 12733, pages 155-171, 2021. Springer. [WWW] [doi:10.1007/978-3-030-77867-5_10]
    Abstract:
    During the last decade, the need for reliable and massive Knowledge Graphs (KG) increased. KGs can be created in several ways: manually with forms or automatically with Information Extraction (IE), a natural language processing task for extracting knowledge from text. Relation Extraction is the part of IE that focuses on identifying relations between named entities in texts, which amounts to find new edges in a KG. Most recent approaches rely on deep learning, achieving state-of-the-art performances. However, those performances are still too low to fully automatize the construction of reliable KGs, and human interaction remains necessary. This is made difficult by the statistical nature of deep learning methods that makes their predictions hardly interpretable. In this paper, we present a new symbolic and interpretable approach for Relation Extraction in texts. It is based on a modeling of the lexical and syntactic structure of text as a knowledge graph, and it exploits extit{Concepts of Neighbours}, a method based on Graph-FCA for computing similarities in knowledge graphs. An evaluation has been performed on a subset of TACRED (a relation extraction benchmark), showing promising results.
    [bibtex-key = AyaCelFer2021icfca] [bibtex-entry]


  6. Francesco Bariatti, Peggy Cellier, and Sébastien Ferré. GraphMDL+: interleaving the generation and MDL-based selection of graph patterns. In Chih-Cheng Hung, Jiman Hong, Alessio Bechini, and Eunjee Song, editors, ACM/SIGAPP Symp. Applied Computing (SAC), pages 355-363, 2021. ACM. [WWW] [doi:10.1145/3412841.3441917]
    Abstract:
    Graph pattern mining algorithms ease graph data analysis by extracting recurring structures. However, classic pattern mining approaches tend to extract too many patterns for human analysis. Recently, the GraphMDL algorithm has been proposed, which reduces the generated pattern set by using the \emph{Minimum Description Length} (MDL) principle to select a small descriptive subset of patterns. The main drawback of this approach is that it needs to first generate all possible patterns and then sieve through their complete set. In this paper we propose GraphMDL+, an approach based on the same description length definitions as GraphMDL but which tightly interleaves pattern generation and pattern selection (instead of generating all frequent patterns beforehand), and outputs a descriptive set of patterns at any time. Experiments show that our approach takes less time to attain equivalent results to GraphMDL and can attain results that GraphMDL could not attain in feasible time. Our approach also allows for more freedom in the pattern and data shapes, since it is not tied to an external approach.
    [bibtex-key = BarCelFer2021sac] [bibtex-entry]


  7. Shridhar B. Dandin and Mireille Ducassé. ComVisMD -- Compact 2D Visualization of Multidimensional Data: Experimenting with Two Different Datasets. In Harish et al. Sharma, editor, Intelligent Learning for Computer Vision, volume 61 of Lecture Notes on Data Engineering and Communications Technologies, pages 473-485, 2021. Springer Singapore. [WWW] Keyword(s): Tabular data, Visual representation design, Data analysis, Reasoning, Problem solving, Decision making, Data clustering, Aggregation.
    Abstract:
    Interpreting data with many attributes is a difficult issue. A simple 2D display, projecting two attributes onto two dimensions, is relatively easy to interpret but provides limited help to see multidimensional correlations. We propose a tool, ComVisMD, which displays, from a dataset, five dimensions in compact 2D maps. A map contains cells; each one represents an object from the dataset. In addition to the usual horizontal and vertical projections and the use of colors, we offer holes and shapes. In order to compact the display, we partition objects according to two dimensions, grouping values of each dimension into up to seven categories. In this paper, we present two case studies covering two different domains, a cricket player dataset and a heart disease dataset. The cricket dataset has 15 attributes and 2170 objects. We show how, using ComVisMD, correlations between variables can be found in an intuitive way. The heart disease dataset has 14 attributes and 297 objects. Blokh and Stambler, in the June 2015 issue of ``Aging and Disease,'' state that individual attributes show little correlation with heart disease. Yet in combination the correlation improves dramatically. We show how ComVisMD helps visualize those multidimensional correlations between four attributes and heart disease diagnosis.
    [bibtex-key = dandin2021] [bibtex-entry]


  8. Sébastien Ferré. Adding Structure and Removing Duplicates in SPARQL Results with Nested Tables. In Further with Knowledge Graphs, pages 227-240, 2021. IOS Press. [doi:10.3233/SSW210047]
    Abstract:
    The results of a SPARQL query are generally presented as a table with one row per result, and one column per projected variable. This is an immediate consequence of the formal definition of SPARQL results as a sequence of mappings from variables to RDF terms. However, because of the flat structure of tables, some of the RDF graph structure is lost. This often leads to duplicates in the contents of the table, and difficulties to read and interpret results. % We propose to use nested tables to improve the presentation of SPARQL results. A nested table is a table where cells may contain embedded tables instead of RDF terms, and so recursively. We introduce an automated procedure that lifts flat tables into nested tables, based on an analysis of the query. We have implemented the procedure on top of Sparklis, a guided query builder in natural language, in order to further improve the readability of its UI. It can as well be implemented on any SPARQL querying interface as it only depends on the query and its flat results. We illustrate our proposal in the domain of pharmacovigilance, and evaluate it on complex queries over Wikidata.
    [bibtex-key = Fer2021semantics] [bibtex-entry]


  9. Sébastien Ferré. Analytical Queries on Vanilla RDF Graphs with a Guided Query Builder Approach. In Troels Andreasen, Guy De Tré, Janusz Kacprzyk, Henrik Legind Larsen, Gloria Bordogna, and Slawomir Zadrozny, editors, Flexible Query Answering Systems, LNCS 12871, pages 41-53, 2021. Springer. [WWW] [doi:10.1007/978-3-030-86967-0_4]
    Abstract:
    As more and more data are available as RDF graphs, the availability of tools for data analytics beyond semantic search becomes a key issue of the Semantic Web. Previous work require the modelling of data cubes on top of RDF graphs. We propose an approach that directly answers analytical queries on unmodified (vanilla) RDF graphs by exploiting the computation features of SPARQL~1.1. We rely on the NAF design pattern to design a query builder that completely hides SPARQL behind a verbalization in natural language; and that gives intermediate results and suggestions at each step. Our evaluations show that our approach covers a large range of use cases, scales well on large datasets, and is easier to use than writing SPARQL queries.
    [bibtex-key = Fer2021fqas] [bibtex-entry]


2020
  1. Peggy Cellier and Kurt Driessens, editors. Machine Learning and Knowledge Discovery in Databases - International Workshops of ECML PKDD 2019, Würzburg, Germany, September 16-20, 2019, Proceedings, Part I, volume 1167 of Communications in Computer and Information Science, 2020. Springer. [bibtex-key = DBLP:conf/pkdd/2019-w1] [bibtex-entry]


  2. Peggy Cellier and Kurt Driessens, editors. Machine Learning and Knowledge Discovery in Databases - International Workshops of ECML PKDD 2019, Würzburg, Germany, September 16-20, 2019, Proceedings, Part II, volume 1168 of Communications in Computer and Information Science, 2020. Springer. [bibtex-key = DBLP:conf/pkdd/2019-w2] [bibtex-entry]


  3. Sébastien Ferré, editor. IC 2020 : 31es Journées francophones d'Ingénierie des Connaissances, Angers, France, June 29 - July 3, 2020, 2020. [bibtex-key = IC2020] [bibtex-entry]


  4. Peggy Cellier. Towards Usable Pattern Mining. PhD thesis, Habilitation à diriger des recherche de l'université de Rennes 1, 30 octobre 2020. [bibtex-key = Cellier2020HDR] [bibtex-entry]


  5. Sébastien Ferré, Mehdi Kaytoue, Marianne Huchard, Sergei O. Kuznetsov, and Amedeo Napoli. A guided tour of artificial intelligence research, volume II, chapter Formal Concept Analysis: from knowledge discovery to knowledge processing (Chapter 13), pages 411-445. Springer, 2020. [WWW] Keyword(s): formal concept analysis, relational concept analysis, Graph-FCA.
    Abstract:
    In this chapter, we introduce Formal Concept Analysis (FCA) and some of its extensions. FCA is a formalism based on lattice theory aimed at data analysis and knowledge processing. FCA allows the design of so-called concept lattices from binary and complex data. These concept lattices provide a realistic basis for knowledge engineering and the design of knowledge-based systems. Indeed, FCA is closely related to knowledge discovery in databases, knowledge representation and reasoning. Accordingly, FCA supports a wide range of complex and intelligent tasks among which classification, information retrieval, recommendation, network analysis, software engineering and data management. Finally, FCA is used in many applications demonstrating its growing importance in data and knowledge sciences.
    [bibtex-key = BookAI:FCA] [bibtex-entry]


  6. Carlos Bobed, Pierre Maillot, Peggy Cellier, and Sébastien Ferré. Data-driven Assessment of Structural Evolution of RDF Graphs. Semantic Web: Interoperability, Usability, Applicability, 11:831-853, 2020. [WWW] Keyword(s): semantic web, data mining, data quality, data evolution, data-driven approach, MDL principle.
    Abstract:
    Since the birth of the Semantic Web, numerous knowledge bases have appeared. The applications that exploit them rely on the quality of their data through time. In this regard, one of the main dimensions of data quality is conformance to the expected usage of the vocabulary. However, the vocabulary usage (i.e., how classes and properties are actually populated) can vary from one base to another. Moreover, through time, such usage can evolve within a base and diverges from the previous practices. Methods have been proposed to follow the evolution of a knowledge base by the observation of the changes of their intentional schema (or ontology); however, they do not capture the evolution of their actual data, which can vary greatly in practice. In this paper, we propose a data-driven approach to assess the global evolution of vocabulary usage in large RDF graphs. Our proposal relies on two structural measures defined at different granularities (dataset vs update), which are based on pattern mining techniques. We have performed a thorough experimentation which shows that our approach is scalable, and can capture structural evolution through time of both synthetic (LUBM) and real knowledge bases (different snapshots and updates of DBpedia).
    [bibtex-key = BobMaiCelFer2019swj] [bibtex-entry]


  7. Sébastien Ferré and Peggy Cellier. Graph-FCA: An extension of formal concept analysis to knowledge graphs. Discrete Applied Mathematics, 273(5):81-102, 2020. [WWW] [doi:https://doi.org/10.1016/j.dam.2019.03.003] Keyword(s): Formal concept analysis, Knowledge graph, Semantic web, Graph homomorphism.
    Abstract:
    Knowledge graphs offer a versatile knowledge representation, and have been studied under different forms, such as conceptual graphs or RDF graphs in the Semantic Web. A challenge is to discover conceptual structures in those graphs, in the same way as Formal Concept Analysis (FCA) discovers conceptual structures in tables. FCA has been successful for analysing, mining, learning, and exploring tabular data, and our aim is to help transpose those results to graph-based data. Previous several FCA approaches have already addressed relational data, hence graphs, but with various limits. We propose Graph-FCA as an extension of FCA where a dataset is a hypergraph instead of a binary table. We show that it can be formalized simply by replacing objects by tuples of objects. This leads to the notion of "n-ary concept", whose extent is an n-ary relation of objects, and whose intent is a "projected graph pattern". In this paper, we formally reconstruct the fundamental results of FCA for knowledge graphs. We describe in detail the representation of hypergraphs, and the operations on them, as they are much more complex than the sets of attributes that they extend. We also propose an algorithm based on a notion of "pattern basis" to generate and display n-ary concepts in a more efficient and more compact way. We explore a few use cases, in order to study the feasibility and usefulness of Graph-FCA. We consider two use cases: workflow patterns in cooking recipes and linguistic structures from parse trees. In addition, we report on experiments about quantitative aspects of the approach.
    [bibtex-key = FerCel2019dam] [bibtex-entry]


  8. Flavien Lécuyer, Valérie Gouranton, Aurélien Lamercerie, Adrien Reuzeau, Bruno Arnaldi, and Benoît Caillaud. Unveiling the implicit knowledge, one scenario at a time. Visual Computer, pp 1-12, 2020. [WWW] [PDF] [doi:10.1007/s00371-020-01904-7] Keyword(s): Process Mining, demodocos, S3PMSunset, ProcMining. [bibtex-key = lecuyer:hal-02879083] [bibtex-entry]


  9. Romaric Marcilly, Laura Douze, Sébastien Ferré, Bissan Audeh, Carlos Bobed, Agnès Lillo-Le-Louët, Jean-Baptiste Lamy, and Cédric Bousquet. How to interact with medical terminologies? Formative usability evaluations comparing three approaches for supporting the use of MedDRA by pharmacovigilance specialists. BMC Medical Informatics and Decision Making, 20(261), 2020. [WWW] [doi:https://doi.org/10.1186/s12911-020-01280-1] [bibtex-key = marcilly2020midm] [bibtex-entry]


  10. Francesco Bariatti, Peggy Cellier, and Sébastien Ferré. GraphMDL : sélection de motifs de graphes avec le principe MDL. In Extraction et Gestion des Connaissances (EGC), Bruxelles, Belgium, 2020. [WWW] [PDF]
    Abstract:
    Plusieurs algorithmes de fouille de motifs ont été proposés pour identifier des structures récurrentes dans les graphes. Le principal défaut de ces approches est qu'elles produisent généralement trop de motifs pour qu'une analyse humaine soit possible. Récemment, des méthodes de fouille de motifs ont traité ce problème sur des données transactionnelles, séquentielles et relationnelles en utilisant le principe MDL (Minimum Description Length). Dans ce papier, nous proposons une approche MDL pour sélectionner un sous-ensemble représentatif de motifs sur des graphes non-orientés étiquetés. Une notion clé de notre approche est l'introduction de ports pour encoder les connections entre occurrences de motifs, sans perte d'information. Nos expériences montrent que le nombre de motifs est drastiquement réduit et que les motifs sélectionnés peuvent avoir des formes complexes.
    [bibtex-key = Bariatti2020egc] [bibtex-entry]


  11. Francesco Bariatti, Peggy Cellier, and Sébastien Ferré. GraphMDL Visualizer: Interactive Visualization of Graph Patterns. In Graph Embedding and Mining (GEM), an ECML-PKDD workshop, 2020. [WWW] Keyword(s): graph mining, graph pattern, MDL principle, visualization.
    Abstract:
    Pattern mining algorithms allow to extract structures from data to highlight interesting and useful knowledge. However, those approaches can only be truly helpful if the users can actually understand their outputs. Thus, visualization techniques play a great role in pattern mining, bridging the gap between the algorithms and the users. In this demo paper we propose GraphMDL Visualizer, a tool for the interactive visualization of the graph patterns extracted with GraphMDL, a graph mining approach based on the MDL principle. GraphMDL Visualizer is structured according to the behavior and needs of users when they analyze GraphMDL results. The tool has different views, ranging from more general (distribution of pattern characteristics), to more specific (visualization of specific patterns). It is also highly interactive, allowing the users to customize the different views, and navigate between them, through simple mouse clicks. GraphMDL Visualizer is freely available online.
    [bibtex-key = BarCelFer2020gem] [bibtex-entry]


  12. Francesco Bariatti, Peggy Cellier, and Sébastien Ferré. GraphMDL: Graph Pattern Selection based on Minimum Description Length. In Symposium on Intelligent Data Analysis (IDA), 2020. [WWW] [PDF] Keyword(s): Minimum Description Length, Graph Mining, Pattern Mining.
    Abstract:
    Many graph pattern mining algorithms have been designed to identify recurring structures in graphs. The main drawback of these approaches is that they often extract too many patterns for human analysis. Recently, pattern mining methods using the Minimum Description Length (MDL) principle have been proposed to select a characteristic subset of patterns from transactional, sequential and relational data. In this paper, we propose an MDL-based approach for selecting a characteristic subset of patterns on labeled graphs. A key notion in this paper is the introduction of ports to encode connections between pattern occurrences without any loss of information. Experiments show that the number of patterns is drastically reduced. The selected patterns have complex shapes and are representative of the data.
    [bibtex-key = Bariatti2020ida] [bibtex-entry]


  13. Mireille Ducassé. Kartu-Verbs: A Semantic Web Base of Inflected Georgian Verb Forms to Bypass Georgian Verb Lemmatization Issues. In Zoe Gavriilidou, editor, Proceedings of XIX EURALEX Conference, 2020-2021. Euralex association. [WWW] Keyword(s): Georgian verbs, Inflected forms, Dictionary front-end, Semantic web tool, Prolog.
    Abstract:
    The Georgian language has a complex verbal system, both agglutinative and inflectional, with many irregularities. Inflected forms of a given verb can differ greatly from one another and it is still a controversial issue to determine which lemmas should represent a verb in dictionaries. Verb tables help people to track lemmas starting from inflected forms but these tables are tedious and error-prone to browse. We propose Kartu-Verbs, a Semantic Web base of inflected Georgian verb forms. For a given verb, all its inflected forms are present. Knowledge can easily be traversed in all directions: from Georgian to French and English; from an inflected form to a masdar (a verbal noun, the form that comes closest to an infinitive), and conversely from a masdar to any inflected form; from component(s) to forms and from a form to its components. Users can easily retrieve the lemmas that are relevant to access their preferred dictionaries. Kartu-Verbs can be seen as a front-end to any Georgian dictionary, thus bypassing the lemmatization issues.
    [bibtex-key = ducasse2020] [bibtex-entry]


  14. Mireille Ducassé. Kartu-Verbs: A Semantic Web Base of Inflected Georgian Verb Forms to Bypass Georgian Verb Lemmatization Issues. In Zoe Gavriilidou, editor, First Proceedings of XIX EURALEX Conference, November 2020. Euralex association. [WWW] Keyword(s): Georgian verbs, Inflected forms, Dictionary front-end, Semantic web tool, Prolog.
    Abstract:
    The Georgian language has a complex verbal system, both agglutinative and inflectional, with many irregularities. Inflected forms of a given verb can differ greatly from one another and it is still a controversial issue to determine which lemmas should represent a verb in dictionaries. Verb tables help people to track lemmas starting from inflected forms but these tables are tedious and error-prone to browse. We propose Kartu-Verbs, a Semantic Web base of inflected Georgian verb forms. For a given verb, all its inflected forms are present. Knowledge can easily be traversed in all directions: from Georgian to French and English; from an inflected form to a masdar (a verbal noun, the form that comes closest to an infinitive), and conversely from a masdar to any inflected form; from component(s) to forms and from a form to its components. Users can easily retrieve the lemmas that are relevant to access their preferred dictionaries. Kartu-Verbs can be seen as a front-end to any Georgian dictionary, thus bypassing the lemmatization issues.
    [bibtex-key = ducasse2020] [bibtex-entry]


  15. Sébastien Ferré. A Proposal for Nested Results in SPARQL. In K. Taylor, R. Gonçalves, F. Lecue, and J. Yan, editors, ISWC 2020 Posters, Demos, and Industry Tracks, volume 2721 of CEUR Workshop Proceedings, pages 114-119, 2020. [WWW] Keyword(s): SPARQL, nested tables.
    Abstract:
    Tables are a common form of query results, notably in SPARQL. However, due to the flat structure of tables, all structure from the RDF graph is lost, and this can lead to duplications in the table contents, and difficulties to interpret the results. We propose an extension of SPARQL 1.1 aggregations to get nested results, i.e. tables where cells may contain embedded tables instead of RDF terms, and so recursively
    [bibtex-key = Fer2020iswc] [bibtex-entry]


  16. Sébastien Ferré. Construction guidée de requêtes analytiques sur des graphes RDF. In Atelier Web des Données, Bruxelles, Belgium, 2020. [WWW] [PDF] Keyword(s): RDF, SPARQL, analytical query, statistical query, Sparklis, query builder.
    Abstract:
    As more and more data are available as RDF graphs, the availability of tools for analytical queries beyond semantic search becomes a key issue of the Semantic Web. Previous work require the modelling of data cubes on top of RDF graphs. We propose an approach that directly answers analytical queries on unmodified RDF graphs by exploiting the computation features of SPARQL 1.1 (aggregations, expressions). We rely on the NAF design pattern to design a query builder user interface that is user-friendly by completely hiding SPARQL behind a verbalization in natural language; and responsive by giving intermediate results and suggestions at each step. Our evaluations show that our approach covers a large range of use cases, and scales well on large datasets.
    [bibtex-key = Fer2020awd] [bibtex-entry]


  17. Nicolas Fouqué, Sébastien Ferré, and Peggy Cellier. Concepts de voisins dans les graphes RDF : Une extension Jena et une interface graphique. In Antoine Cornuéjols and Etienne Cuvelier, editors, Extraction et Gestion des Connaissances (EGC), volume E-36 of RNTI, pages 483-490, 2020. Éditions RNTI. [WWW] Keyword(s): RDF, concepts de voisins.
    Abstract:
    Les concepts de voisins définissent une forme symbolique de similarité entre les entités d'un graphe de connaissances. Partant d'une entité, chaque concept de voisins est un cluster d'entités voisines partageant un même motif de graphe centré sur l'entité. Dans ce papier démo, nous rappelons les définitions des concepts de voisins et nous présentons une extension de la librairie Jena dont l'API permet de calculer les concepts de voisins pour un modèle RDF(S) Jena. Nous présentons également une interface graphique permettant à un utilisateur d'effectuer ces calculs de façon simple et interactive.
    [bibtex-key = Fouque2020egc] [bibtex-entry]


  18. Clément Gautrais, Peggy Cellier, Matthijs van Leeuwen, and Alexandre Termier. Widening for MDL-Based Retail Signature Discovery. In Michael R. Berthold, Ad Feelders, and Georg Krempl, editors, Advances in Intelligent Data Analysis XVIII - 18th International Symposium on Intelligent Data Analysis, IDA 2020, Konstanz, Germany, April 27-29, 2020, Proceedings, volume 12080 of Lecture Notes in Computer Science, pages 197-209, 2020. Springer. [bibtex-key = IDA20Gautrais] [bibtex-entry]


  19. Priscilla Keip, Sébastien Ferré, Alain Gutierrez, Marianne Huchard, Pierre Silvie, and Pierre Martin. Practical Comparison of FCA Extensions to Model Indeterminate Value of Ternary Data. In F. J. Valverde-Albacete and M. Trnecka, editors, Int. Conf. Concept Lattices and Their Applications, volume 2668 of CEUR Workshop Proceedings, pages 197-208, 2020. CEUR-WS.org. [WWW] Keyword(s): FCA, Graph-FCA, RCA, triadic FCA, relational data.
    Abstract:
    The Knomana knowledge base brings together knowledge from the scientific literature on the use of plants with pesticidal or antibiotic effects on animals, plants, and human beings to propose protection solutions using local plants. In this literature, the elements of the 3-tuple (protected organism, protecting plant, pest) are named using the binomial nomenclature consisting of the genus name followed by the species name. In some instances, authors use the abbreviation "sp." in the singular or "spp." in the plural, as species name, to indicate the indeterminate status of the species for a guaranteed genus. To suggest protection solutions, the indeterminacy of the species has to be hypothesized based on assigning the sp./spp. to the other species in the same genus and conversely. This paper discusses the classification of ternary data containing some indeterminate values generated by three extensions of Formal Concept Analysis.
    [bibtex-key = Keip2020cla] [bibtex-entry]


  20. Aurélien Lamercerie. ARES : un extracteur d'exigences pour la modélisation de systèmes. In EGC 2020 - Extraction et Gestion des Connaissances (Atelier - Fouille de Textes - Text Mine), Bruxelles, Belgium, pages 1-4, January 2020. [WWW] [PDF] [bibtex-key = lamercerie:hal-02971727] [bibtex-entry]


  21. Aurélien Lamercerie. Transduction sémantique pour la modélisation de système. In PFIA 2020 - Plate-Forme de l'Intelligence Artificielle (PFIA), rencontres RJCIA, Angers, France, pages 1-6, June 2020. [WWW] [PDF] Keyword(s): System Design, semantic parsing, abstract representation, formalization. [bibtex-key = lamercerie:hal-02971742] [bibtex-entry]


  22. Aurélien Lamercerie and Benoît Caillaud. An Algebra of Deterministic Propositional Acceptance Automata (DPAA). In FDL 2020 - Forum on specification & Design Languages, Kiel, Germany, pages 1-8, September 2020. [WWW] [PDF] Keyword(s): Interface Theory, Automata for System Analysis, Discrete Time Reactive System, Requirements Engineering. [bibtex-key = lamercerie:hal-02971772] [bibtex-entry]


2019
  1. Annie Foret and Denis Béchet. On Categorial Grammatical Inference and Logical Information Systems. In Logic and Algorithms in Computational Linguistics 2018, Series: Advances in Intelligent Systems and Computing, Studies in Computational Intelligence. 2019. [WWW]
    Abstract:
    In this chapter, learning is viewed as a symbolic issue in an unsupervised setting,from raw or from structured data, for some variants of Lambek grammars and of categorial dependency grammars. For these frameworks, the authors present different type connectives and structures, some limitations, and some algorithms. On the experimental side, categorial grammar has potentials as a particular case of Logical Information System.
    [bibtex-key = foret:hal-02462675] [bibtex-entry]


  2. Alex Chauvet and Annie Foret. Systèmes d'information pour les chercheurs en droit. In Colloque des Convergences du Droit et du Numérique, Actes du colloque des Convergences du droit et du numérique, Bordeaux, France, July 2019. Université de Bordeaux. [WWW] [PDF] Keyword(s): Droit, Numérique, Convergences. [bibtex-key = foret:2019a] [bibtex-entry]


  3. Sébastien Ferré. Link Prediction in Knowledge Graphs with Concepts of Nearest Neighbours. In Pascal Hitzler, Miriam Fernández, Krzysztof Janowicz, Amrapali Zaveri, Alasdair J. G. Gray, Vanessa López, Armin Haller, and Karl Hammar, editors, The Semantic Web (ESWC), LNCS 11503, pages 84-100, 2019. Springer. [WWW] [doi:10.1007/978-3-030-21348-0_6] Keyword(s): semantic web, link prediction, nearest neighbours, similarity, graph pattern, concept. [bibtex-key = Fer2019eswc] [bibtex-entry]


  4. Sébastien Ferré. Prédiction de liens dans les graphes de connaissances avec les concepts de plus proches voisins. In Marie-Christine Rousset and Lydia Boudjeloud-Assala, editors, Extraction et Gestion des connaissances (EGC), volume E-35 of RNTI, pages 69-80, 2019. Hermann-Éditions. [WWW] Keyword(s): semantic web, link prediction, nearest neighbours, similarity, graph pattern, concept. [bibtex-key = Fer2019egc] [bibtex-entry]


  5. Mireille Ducassé. Kartu-Verbs: A Semantic Web Base of Inflected Verb Forms for the Georgian Language. Technical Report, IRISA-INSA, December 2019. Note: Avalaible under HAL https://hal.archives-ouvertes.fr/hal-02460975. [WWW]
    Abstract:
    The Georgian language has a complex verbal system, both agglutinative and inflectional, with many exceptions. It is still a controversial issue to deter- mine which lemmas should represent a verb in dictionaries. Verb tables help neophytes to track lemmas starting from inflected forms but if in paper documents they are tedious and error-prone to browse. We propose Kartu-Verbs, a Semantic Web base of inflected Georgian verb forms. For a given verb, all inflected forms are present. Knowledge can easily be traversed in all directions: from Georgian to French and English; from an inflected form to a masdar, and conversely from a masdar to any inflected form; from component(s) to forms and from a form to its components. Users can easily retrieve the lemmas that are relevant to access their preferred dictionary. Kartu-Verbs can be seen as a front-end to any Georgian dictionary, thus bypassing the lemmatization issues. This report illustrates in detail how to use Kartu-Verbs and gives some indications about how the base is built. Our base, in its current state, is already a successful proof of concept. It has proven helpful to learn about Georgian verbs.
    [bibtex-key = ducasse2019] [bibtex-entry]


  6. Annie Foret. ISNLP on AI for Education - Exploratory action 2019 : Exploratory experiments combining Information Systems and Natural Language Processing. Cominlabs days, November 2019. Note: Poster. [WWW] [PDF] [bibtex-key = foret:hal-02462991] [bibtex-entry]


  7. Aurélien Lamercerie. Une algèbre des automates d'acceptation propositionnelle déterministes comme théorie d'interface pour la conception de systèmes cyberphysiques. MSR 2019 - 12ème Colloque sur la Modélisation des Systèmes Réactifs, Nov 2019, Angers, France, November 2019. Note: Poster. [WWW] [PDF] [bibtex-key = lamercerie:hal-02432696] [bibtex-entry]


2018
  1. Annie Foret, Gregory M. Kobele, and Sylvain Pogodalla, editors. Formal Grammar 2018 - 23rd International Conference, FG 2018, Sofia, Bulgaria, August 11-12, 2018, Proceedings, volume 10950 of Lecture Notes in Computer Science, 2018. Springer. [WWW] [doi:10.1007/978-3-662-57784-4] [bibtex-key = FG2018] [bibtex-entry]


  2. Annie Foret, Reinhard Muskens, and Sylvain Pogodalla, editors. Formal Grammar - 22nd International Conference, FG 2017, Toulouse, France, July 22-23, 2017, Revised Selected Papers, volume 10686 of Lecture Notes in Computer Science, 2018. Springer. [WWW] [doi:10.1007/978-3-662-56343-4] [bibtex-key = DBLP:conf/fg/2017] [bibtex-entry]


  3. Anne-Laure Ligozat, Peggy Cellier, Anne-Lyse Minard, Vincent Claveau, Cyril Grouin, and Patrick Paroubek, editors. Actes de la conférence Traitement Automatique de la Langue Naturelle, TALN 2018, 2018. [WWW] [PDF] [bibtex-key = TALN2018] [bibtex-entry]


  4. Pegah Alizadeh, Peggy Cellier, Thierry Charnois, Bruno Crémilleux, and Albrecht Zimmermann. An Experimental Approach For Information Extraction in Multi-Party Dialogue Discourse. In Proceedings of the 19th International Conference on Computational Linguistics and Intelligent Text Processing (CiCling 2018), LNCS, 2018. Springer. [WWW] [bibtex-key = AlCelCh2018cicling] [bibtex-entry]


  5. Pegah Alizadeh, Peggy Cellier, Thierry Charnois, Bruno Crémilleux, and Albrecht Zimmermann. Étude Expérimentale d'Extraction d'Information dans des Retranscriptions de Réunions. In Traitement automatique du langage naturel (TALN), Rennes, France, May 2018. Keyword(s): Topic Modeling, Information Extraction, Dialogue Texts. [bibtex-key = alizadeh2018TALN] [bibtex-entry]


  6. Carlos Bobed, Laura Douze, Sébastien Ferré, and Romaric Marcilly. PEGASE: A Knowledge Graph for Search and Exploration in Pharmacovigilance Data. In P. Cimiano and O. Corby, editors, EKAW Posters and Demonstrations, volume 2262 of CEUR Workshop Proceedings, pages 33-36, 2018. [WWW] [PDF] Keyword(s): PEGASE, Sparklis, knowledge graph.
    Abstract:
    Pharmacovigilance is in charge of studying the adverse effects of pharmaceutical products. In this field, pharmacovigilance specialists experience several difficulties when searching and exploring their patient data despite the existence of standardized terminologies (MedDRA). In this paper, we present our approach to enhance the way pharmacovigilance specialists perform search and exploration on their data. First, we have developed a knowledge graph that relies on the OntoADR ontology to semantically enrich the MedDRA terminology with SNOMED CT concepts, and that includes anonymized patient data from FAERS. Second, we have chosen and applied a semantic search tool, Sparklis, according to the user requirements that we have identified in pharmacovigilance.
    [bibtex-key = Pegase2018ekaw] [bibtex-entry]


  7. Carlos Bobed, Laura Douze, Sébastien Ferré, and Romaric Marcilly. Sparklis over PEGASE knowledge graph: a new tool for pharmacovigilance. In A. Waagmeester, C. J. O. Baker, A. Splendiani, O. Deniz Beyan, and M. Scott Marshall, editors, Int. Conf. Semantic Web Applications and Tools for Life Sciences (SWAT4LS), volume 2275 of CEUR Workshop Proceedings, 2018. [WWW] [PDF] Keyword(s): PEGASE, Sparklis, knowledge graph.
    Abstract:
    Pharmacovigilance is in charge of studying the adverse effects of pharmaceutical products. In this field, pharmacovigilance specialists experience several difficulties when searching and exploring their patient data despite the existence of standardized terminologies (MedDRA). In this paper, we present our approach to enhance the way pharmacovigilance specialists perform search and exploration on their data. First, we have developed a knowledge graph that relies on the OntoADR ontology to semantically enrich the MedDRA terminology with SNOMED CT concepts, and that includes anonymized patient data from FAERS. Second, we have chosen and extended a semantic search tool, Sparklis, according to the user requirements that we have identified in pharmacovigilance. We report the results of a usability evaluation that has been performed by human factors specialists to check the benefits of our proposal.
    [bibtex-key = Pegase2018swat] [bibtex-entry]


  8. Shridhar Dandin and Mireille Ducassé. ComVisMD-A Visualization Tool for Compact Display of Multidimensional Data: An illustration on Cricket Players Data. In IEEE VIS 2018 International Conference, Abstract, 2018. Note: Video previews available under https://vimeo.com/groups/vis18previews. [WWW]
    Abstract:
    Database information is multidimensional and often displayed in tabular format (row/column display). A Choropleth map is a thematic map in which areas are colored according to a variable of interest. They are used mostly for compact graphical representation of geographical information. We propose a system, ComVisMD, inspired by choropleth map, to visualize multidimensional data taking sets of 4 dimensions and projecting them on a compact 2D-display. The first dimension uses the attribute of main interest to color areas according to a 5-color scale. The next 2 dimensions define the displayed areas as square cells and give the horizontal and vertical axes. The fourth dimension is displayed in the form of varying-size holes in the cells. We illustrate our approach on cricket players data and show how ComVisMDs compact visualization can help analyze data and find correlations as well as explain the exceptions by the way of intuitive color observation, shape of the cells, information on cell, dynamic scaling, classification and clustering.
    [bibtex-key = Dandin2018] [bibtex-entry]


  9. Sébastien Ferré. Answers Partitioning and Lazy Joins for Efficient Query Relaxation and Application to Similarity Search. In A. Gangemi and others, editors, Int. Conf. The Semantic Web (ESWC), LNCS 10843, pages 209-224, 2018. Springer. [WWW] [doi:10.1007/978-3-319-93417-4_14] Keyword(s): query relaxation, approximate answer, similarity search, semantic similarity, TDF, graph pattern, partition, join.
    Abstract:
    Query relaxation has been studied as a way to find approximate answers when user queries are too specific or do not align well with the data schema. We are here interested in the application of query relaxation to similarity search of RDF nodes based on their description. However, this is challenging because existing approaches have a complexity that grows in a combinatorial way with the size of the query and the number of relaxation steps. We introduce two algorithms, answers partitioning and lazy join, that together significantly improve the efficiency of query relaxation. Our experiments show that our approach scales much better with the size of queries and the number of relaxation steps, to the point where it becomes possible to relax large node descriptions in order to find similar nodes. Moreover, the relaxed descriptions provide explanations for their semantic similarity.
    [bibtex-key = Fer2018eswc] [bibtex-entry]


  10. Sébastien Ferré. Responsive and Flexible Controlled Natural Language Authoring with Zipper-Based Transformations. In B. Davis, C. M. Keet, and A. Wyner, editors, Int. Work. Controlled Natural Language (CNL), volume 304 of Frontiers in Artificial Intelligence and Applications, pages 21-30, 2018. IOS Press. [WWW] [doi:10.3233/978-1-61499-904-1-21] Keyword(s): controlled natural language, authoring, user interaction, abstract syntax, zipper, focus.
    Abstract:
    Controlled natural languages (CNL) have the benefits to combine the readability of natural languages, and the accuracy of formal languages. They have been used to help users express facts, rules or queries. While generally easy to read, CNLs remain difficult to write because of the constrained syntax. A common solution is a grammar-based auto-completion mechanism to suggest the next possible words in a sentence. However, this solution has two limitations: (a) partial sentences may have no semantics, which prevents giving intermediate results or feedback, and (b) the suggestion is often limited to adding words at the end of the sentence. % We propose a more responsive and flexible CNL authoring by designing it as a sequence of sentence transformations. Responsiveness is obtained by having a complete, and hence interpretable, sentence at each time. Flexibility is obtained by allowing insertion and deletion on any part of the sentence. Technically, this is realized by working directly on the abstract syntax, rather than on the concrete syntax, and by using Huet's zippers to manage the focus on a query part, the equivalent of the text cursor of a word processor.
    [bibtex-key = Fer2018cnl] [bibtex-entry]


  11. Sébastien Ferré. What's New in SPARKLIS. In P. Cimiano and O. Corby, editors, EKAW Posters and Demonstrations, volume 2262 of CEUR Workshop Proceedings, pages 41-44, 2018. [WWW] [PDF] Keyword(s): Sparklis, demo.
    Abstract:
    Sparklis is a SPARQL query builder that can connect to any endpoint, and that interacts with users in natural language only. Users are guided in the building of their queries so that they do not have to know the schema, and so that empty results are almost completely avoided. This demo paper presents a number of recent extensions to Sparklis. Most notably, it now supports analytical queries, Wikidata statement qualifiers, and the display of results on a map or as a slideshow.
    [bibtex-key = Fer2018ekaw] [bibtex-entry]


  12. Sébastien Ferré and Peggy Cellier. How Hierarchies of Concept Graphs Can Facilitate the Interpretation of RCA Lattices?. In D. I. Ignatov and L. Nourine, editors, Int. Conf. Concept Lattices and Their Applications (CLA), CEUR 2123, pages 69-80, 2018. CEUR-WS.org. [WWW] [PDF] Keyword(s): formal concept analysis, relational concept analysis, data mining, concept graph.
    Abstract:
    Relational Concept Analysis (RCA) has been introduced in order to allow concept analysis on multi-relational data. It significantly widens the field of application of Formal Concept Analysis (FCA), and it produces richer concept intents that are similar to concept definitions in Description Logics (DL). However, reading and interpreting RCA concept lattices is notoriously difficult. Nica {\em et al} have proposed to represent RCA intents by cpo-patterns in the special case of sequence structures. We propose an equivalent representation of a family of RCA concept lattices in the form of a hierarchy of concept graphs. Each concept belongs to one concept graph, and each concept graph exhibits the relationships between several concepts. A concept graph is generally transversal to several lattices, and therefore highlights the relationships between different types of objects. We show the benefits of our approach on several use cases from the RCA litterature.
    [bibtex-key = FerCel2018cla] [bibtex-entry]


  13. Esther Galbrun, Peggy Cellier, Nikolaj Tatti, Alexandre Termier, and Bruno Crémilleux. Mining Periodic Patterns with a MDL Criterion. In European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD), 2018. [WWW]
    Abstract:
    The quantity of event logs available is increasing rapidly, be they produced by industrial processes, computing systems, or life tracking , for instance. It is thus important to design effective ways to uncover the information they contain. Because event logs often record repetitive phenomena, mining periodic patterns is especially relevant when considering such data. Indeed, capturing such regularities is instrumental in providing condensed representations of the event sequences. We present an approach for mining periodic patterns from event logs while relying on a Minimum Description Length (MDL) criterion to evaluate candidate patterns. Our goal is to extract a set of patterns that suitably characterises the periodic structure present in the data. We evaluate the interest of our approach on several real-world event log datasets.
    [bibtex-key = GalCel2018ecmlpkdd] [bibtex-entry]


  14. Aurélien Lamercerie. Formal analysis of natural language requirements for the design of cyber-physical systems. In Conférence TALN, 2018. [WWW] [PDF] Keyword(s): Grammatical Formalism, Semantic Representation, Categorial Grammar, Require-ments Engineering, Cyber-physical System, Modal Specification.
    Abstract:
    This paper focuses on the construction of formal representations of natural language texts. The mapping from a natural language to a logical representation is realized with a grammatical formalism, linking the syntactic analysis of the text to a semantic representation. We target the behavioral aspect of the specifications for cyber-physical systems, ie any type of system in which software components interact closely with a physical environment. In this way, the challenge would be to provide assistance to the designer. So, we could simulate and verify, by automatic or assisted methods, "systems" specifications expressed in natural language. This paper presents some existing contributions that could enable progress on this issue.
    [bibtex-key = Lamercerie2018taln] [bibtex-entry]


  15. Son T. Mai, Sihem Amer-Yahia, Ahlame Douzal Chouakria, Ky T. Nguyen, and Anh-Duong Nguyen. Scalable Active Constrained Clustering for Temporal Data. In Int. Conf. Database Systems for Advanced Applications (DASFAA), pages 566-582, 2018. [bibtex-key = Nguyen2018asfaa] [bibtex-entry]


  16. Pierre Maillot and Carlos Bobed. Measuring Structural Similarity Between RDF Graphs. In Proc. of the 33rd ACM/SIGAPP Symposium On Applied Computing (SAC 2018), SWA track, Pau (France), April 2018. ACM. [bibtex-key = SAC2018] [bibtex-entry]


  17. Guillaume Aucher, Anthony Baire, Jean Berbinau, Annie Foret, Jean-Baptiste Lenhof, Marie-Laure Morin, Olivier Ridoux, and François Schwarzentruber. Principles for a Judgement Editor Based on BDD. Research Report, Université de Rennes 1, France, November 2018. Note: Report of a collaborative project between the Cour de cassation, the University of Rennes 1, the ENS Rennes, the INRIA and the University of Torino, from 2013 to 2015.[WWW] [PDF] Keyword(s): BDD, software assistant, Law, artificial intelligence. [bibtex-key = aucher:hal-01914593] [bibtex-entry]


  18. Peggy Cellier, Mireille Ducassé, Sébastien Ferré, and Olivier Ridoux. Data Mining for Fault Localization: towards a Global Debugging Process. Research Report, INSA RENNES ; Univ Rennes, CNRS, IRISA, France, 2018. [WWW] [PDF] [bibtex-key = cellier:hal-02003069] [bibtex-entry]


  19. Hugues Ferreboeuf, Olivier Ridoux, and others. LEAN ICT, Pour une sobriété numérique. Technical report, The SHIFT Project, France, Mars 2018. Note: Rapport de travail intermédiaire.[WWW] [PDF] Keyword(s): Green IT, IT impact, LCA. [bibtex-key = leanICT:2018] [bibtex-entry]


  20. Sébastien Ferré. A SPARQL 1.1 Query Builder for the Data Analytics of Vanilla RDF Graphs. Research Report, IRISA, team SemLIS, 2018. [WWW] [PDF] Keyword(s): RDF, SPARQL, OLAP, natural language interface, data analytics, query builder, expressivity.
    Abstract:
    As more and more data are available as RDF graphs, the availability of tools for data analytics beyond semantic search becomes a key issue of the Semantic Web. Previous work has focused on adapting OLAP-like approaches and question answering by modelling RDF data cubes on top of RDF graphs. We propose a more direct -- and more expressive -- approach by guiding users in the incremental building of SPARQL~1.1 queries that combine several computation features (aggregations, expressions, bindings and filters), and by evaluating those queries on unmodified (vanilla) RDF graphs. We rely on the NF design pattern to hide SPARQL behind a natural language interface, and to provide results and suggestions at every step. We have implemented our approach on top of {\sc Sparklis}, and we report on three experiments to assess its expressivity, usability, and scalability.
    [bibtex-key = Fer2018analytics] [bibtex-entry]


  21. Esther Galbrun, Peggy Cellier, Nikolaj Tatti, Alexandre Termier, and Bruno Crémilleux. Mining Periodic Patterns with a MDL Criterion. Technical report, 2018. [WWW] [bibtex-key = galbrun08report] [bibtex-entry]


2017
  1. Karell Bertet, Daniel Borchmann, Peggy Cellier, and Sébastien Ferré, editors. Formal Concept Analysis: 14th International Conference, ICFCA 2017, Rennes, France, June 13-16, 2017, Proceedings, LNAI10308, 2017. Springer. [doi:10.1007/978-3-319-59271-8] Keyword(s): formal concept analysis. [bibtex-key = ICFCA2017] [bibtex-entry]


  2. Peggy Cellier, Thierry Charnois, Andreas Hotho, Stan Matwin, Marie-Francine Moens, and Yannick Toussaint, editors. Proceedings of the Workshop on Interactions between Data Mining and Natural Language Processing, DMNLP 2017, co-located with the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML-PKDD 2017, Skopje, Macedonia, September 22, 2017, volume 1880 of CEUR Workshop Proceedings, 2017. CEUR-WS.org. [WWW] [bibtex-key = 2017dmnlp] [bibtex-entry]


  3. Sébastien Ferré. Sparklis: An Expressive Query Builder for SPARQL Endpoints with Guidance in Natural Language. Semantic Web: Interoperability, Usability, Applicability, 8(3):405-418, 2017. [WWW] [doi:10.3233/SW-150208] Keyword(s): semantic search, SPARQL endpoint, query builder, faceted search, natural language.
    Abstract:
    Sparklis is a Semantic Web tool that helps users explore and query SPARQL endpoints by guiding them in the interactive building of questions and answers, from simple ones to complex ones. It combines the fine-grained guidance of faceted search, most of the expressivity of SPARQL, and the readability of (controlled) natural languages. No knowledge of the vocabulary and schema are required for users. Many SPARQL features are covered: multidimensional queries, union, negation, optional, filters, aggregations, ordering. Queries are verbalized in either English or French, so that no knowledge of SPARQL is ever necessary. All of this is implemented in a portable Web application, Sparklis, and has been evaluated on many endpoints and questions. No endpoint-specific configuration is necessary as the data schema is discovered on the fly by the tool. Online since April 2014, thousands of queries have been formed by hundreds of users over more than a hundred endpoints.
    [bibtex-key = Fer2016swj] [bibtex-entry]


  4. Guillaume Aucher, Annie Foret, JeanBaptiste Lenhof, Olivier Ridoux, Francois Schwarzentruber, and al.. Principles for a judgement editor based on Multi-BDDs. In MIning and REasoning with Legal texts at ICAIL, London, 2017. [WWW]
    Abstract:
    We describe the theoretical principles that underlie the design of a software tool which could be used by judges for writing judgements and for making decisions about litigations. The tool is based on Binary Decision Diagrams (BDD), which are graphical representations of truth-valued functions associated to propositional formulas. Given a specific litigation, the tool asks questions to the judge; each question is represented by a propositional atom. Their answers, true or false, allow to evaluate the truth value of the formula which encodes the overall recommendation of the software about the litigation. Our approach combines some sort of `theoretical' or `legal' reasoning dealing with the core of the litigation itself together with some sort of `procedural' reasoning dealing with the protocol that has to be followed by the judge during the trial: some questions or group of questions must necessarily be examined and sometimes in a specific order. That is why we consider extensions of BDDs called Multi-BDDs. They are BDDs with multiple entries corresponding to the different specific issues that must necessarily be addressed by the judge during the trial. We illustrate our ideas on a case study dealing with French union trade elections, an example that has been used throughout a project with the French Cour de cassation. We end the article by sketching the architecture of a prototype software that has been developed during this project.
    [bibtex-key = Foret17b] [bibtex-entry]


  5. Alexis Chauvet and Annie Foret. Systèmes d'information pour les chercheurs en droit. In Colloque CDN, Convergences du droit et du numérique, Bordeaux, 2017. [WWW]
    Abstract:
    (fr) Cette étude concerne les systèmes d'information en droit et propose un nouveau prototype. Nous discutons les systèmes d'information actuellement utilisés par les chercheurs en droit, en précisant leurs limites. Nous présentons les principes d'un nouveau prototype pour un meilleur système. Ce travail s'accompagne d'une première réalisation concrète, un système à facettes sémantiques, résultat de notre chaîne de traitement sur un ensemble de décisions du Conseil Constitutionnel. (en) This study concerns information systems in law and proposes a new prototype. We discuss the information systems currently used by legal researchers, including their limitations. We present the principles of a new prototype for a better system. This work is accompanied by a first concrete realization, a system with semantic facets, result of our processing chain on a set of decisions of the Constitutional Council.
    [bibtex-key = Foret17d] [bibtex-entry]


  6. Fabien Chevalier and Sébastien Ferré. Entering the Digital Customer Onboarding Era: How the Semantic Web Can Help. In N. Nikitina, D. Song, A. Fokoue, and P. Haase, editors, Posters & Demonstrations and Industry Tracks at Int. Semantic Web Conf., volume 1963 of CEUR Workshop Proceedings, 2017. CEUR-WS.org. [PDF] Keyword(s): semantic web. [bibtex-key = ChevalierFerre2017iswc] [bibtex-entry]


  7. Shridhar B. Dandin and Mireille Ducassé. Compact Visualization of Database Tables: Experimenting with Cricket Players Data. In 3rd International Conference on Communication System, Pilani, India, Short paper, 2017.
    Abstract:
    Database information is in the form of tables, that is two-dimensional row/column display. For large databases, the tabular data is difficult to read at a time for the aggregated results. Online Analytical database Processing (OLAP) proposes mechanisms to display data in aggregated forms. A choropleth map is a thematic map in which areas are colored in proportion to the measurement of a statistical variable being displayed, such as population density. They are used mostly for compact graphical representation of geographical information. We propose a system that adapts choropleth maps and the OLAP cube to visualize tabular data in a compact way. Our proposal displays multidimensional data like OLAP Cube (a multidimensional data set also called as hyper cube), where we are mapping coloring of database table record based on two attributes $a$ (first dimension) and $b$ (second dimension), mapping varying-size circles based on attribute $c$ (third dimension), mapping numbers based on attribute $d$ (fourth dimension). We illustrate our approach on Cricket players data, namely on two linked tables 'Country' and 'Player'. They have a large number of rows and columns like 16 rows, 9 columns and 251 rows and 17 columns. The visualization presented by the system reduces the size of the table by a factor of about 4, allowing users to grasp more information at a time than the bare table display.
    [bibtex-key = Dandin2017] [bibtex-entry]


  8. Sébastien Ferré. Concepts de plus proches voisins dans des graphes de connaissances. In 28es Journées francophones d'Ingénierie des Connaissances (IC), pages 163-174, 2017. [WWW] [PDF] Keyword(s): semantic web, knowledge graph, nearest neighbours, concept.
    Abstract:
    Nous introduisons la notion de {\em concept de voisins} comme alternative à la notion de distance numérique dans le but d'identifier les objets les plus similaires à un objet requête, comme dans la méthode des plus proches voisins. Chaque concept de voisins est composé d'une intension qui décrit symboliquement ce que deux objets ont en commun et d'une extension qui englobe les objets qui se trouvent entre les deux. Nous définissons ces concepts de voisins pour des données complexes, les graphes de connaissances, où les n{\oe}uds jouent le rôle d'objets. Nous décrivons un algorithme {\em anytime} permettant d'affronter la complexité élevée de la tâche et nous présentons de premières expérimentations sur un graphe RDF de plus de 120.000 triplets.
    [bibtex-key = Ferre2017ic] [bibtex-entry]


  9. Annie Foret. Traitement automatique des langues, données légales, systèmes d'information et logique. In Atelier CDN, Convergences du droit et du numérique, Bordeaux, 2017. [WWW]
    Abstract:
    (fr) Nous proposons d'exploiter les données en utilisant les méthodes de l'analyse de concepts logiques [Ferré et Ridoux (2003)] via des outils de l'équipe LIS : Camelis http://www.irisa.fr/LIS/ferre/camelis/ ou Sparklis http://www.irisa.fr/LIS/ferre/sparklis/. Des expérimentations avec le système de gestion de contexte et le contexte obtenu permettront d'améliorer l'approche et la construction des facettes sémantiques à mettre en avant. (en) We propose to exploit the data using methods of logical concept analysis [Ferré and Ridoux (2003)] using LIS team tools: Camelis http://www.irisa.fr/LIS/ferre/camelis/ or Sparklis http://www.irisa.fr/LIS/ferre/sparklis/. Experiments with the context management system and the context obtained will allow to improve the approach and the construction of semantic facets to be put forward.
    [bibtex-key = Foret17a] [bibtex-entry]


  10. Clément Gautrais, Peggy Cellier, René Quiniou, and Alexandre Termier. Topic Signatures in Political Campaign Speeches. In , pages 2342-2347, 2017. [WWW] [doi:10.18653/v1/D17-1249] [bibtex-key = gautrais_topic_2017] [bibtex-entry]


  11. Clément Gautrais, René Quiniou, Peggy Cellier, Thomas Guyet, and Alexandre Termier. Purchase Signatures of Retail Customers. In Jinho Kim, Kyuseok Shim, Longbing Cao, Jae-Gil Lee, Xuemin Lin, and Yang-Sae Moon, editors, Advances in Knowledge Discovery and Data Mining, Cham, pages 110-121, 2017. Springer International Publishing. [WWW]
    Abstract:
    In the retail context, there is an increasing need for understanding individual customer behavior in order to personalize marketing actions. We propose the novel concept of customer signature, that identifies a set of important products that the customer refills regularly. Both the set of products and the refilling time periods give new insights on the customer behavior. Our approach is inspired by methods from the domain of sequence segmentation, thus benefiting from efficient exact and approximate algorithms. Experiments on a real massive retail dataset show the interest of the signatures for understanding individual customers.
    [bibtex-key = PAKDD17] [bibtex-entry]


  12. Pierre Maillot, Sébastien Ferré, Peggy Cellier, Mireille Ducassé, and Franck Partouche. Nested Forms with Dynamic Suggestions for Quality RDF Authoring. In Int. Conf. Database and Expert Systems Applications (DEXA), pages 35-45, 2017. [WWW] [doi:10.1007/978-3-319-64468-4_3] Keyword(s): semantic web, RDF, knowledge acquisition, form, data quality. [bibtex-key = Formulis17dexa] [bibtex-entry]


  13. Annie Foret. Logical Information Systems for Linguistic Data: TermLis, 2017. [WWW] [PDF]
    Abstract:
    Logical Concept Analysis [1] (LCA) extends the Formal Concept Analysis [4] (FCA) approach. Recently, this approach has been undertaken for terminology, a workflow has been developed to go from XML data to a logical information context. Through experiments on specific resources, facet designs have been tuned to facilitate the search and control on the data. We will consider several usages of such contexts and illustrate benefits of the approach.
    [bibtex-key = Foret17c] [bibtex-entry]


  14. Hedi-Théo Sahraoui, Pierre Holat, Peggy Cellier, Thierry Charnois, and Sébastien Ferré. Exploration of Textual Sequential Patterns, 2017. [bibtex-key = icfcademo17] [bibtex-entry]


  15. Sébastien Ferré. Sparklis: An Expressive Query Builder for SPARQL Endpoints with Guidance in Natural Language. Note: Poster and demo at SEMANTiCS 2017, Amsterdam, 2017. [bibtex-key = Fer2017semantics] [bibtex-entry]


2016
  1. Peggy Cellier, Thierry Charnois, Andreas Hotho, Stan Matwin, Marie-Francine Moens, and Yannick Toussaint, editors. Proceedings of the Workshop on Interactions between Data Mining and Natural Language Processing, DMNLP 2016, co-located with the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML-PKDD 2016, Riva del Garda, Italy, September 23, 2016, volume 1646 of CEUR Workshop Proceedings, 2016. CEUR-WS.org. [WWW] [bibtex-key = DBLP:conf/pkdd/2016dmnlp] [bibtex-entry]


  2. Mouhamadou Ba, Sébastien Ferré, and Mireille Ducassé. Solving Data Mismatches in Bioinformatics Workflows by Generating Data Converters. Transactions on Large-Scale Data and Knowledge-Centered Systems (TLDKS), LNCS 9510:88-115, 2016. Keyword(s): workflow, data conversion, bioinformatics, type system.
    Abstract:
    Heterogeneity of data and data formats in bioinformatics entail mismatches between inputs and outputs of different services, making it difficult to compose them into workflows. To reduce those mismatches, bioinformatics platforms propose ad'hoc converters, called shims. When shims are written by hand, they are time-consuming to develop, and cannot anticipate all needs. When shims are automatically generated, they miss transformations, for example data composition from multiple parts, or parallel conversion of list elements. This article proposes to systematically detect convertibility from output types to input types. Convertibility detection relies on a rule system based on abstract types, close to XML Schema. Types allow to abstract data while precisely accounting for their composite structure. Detection is accompanied by an automatic generation of converters between input and output XML data. % We show the applicability of our approach by abstracting concrete bioinformatics types (e.g., complex biosequences) for a number of bioinformatics services (e.g., blast). We illustrate how our automatically generated converters help to resolve data mismatches when composing workflows. % We conducted an experiment on bioinformatics services and datatypes, using an implementation of our approach, as well as a survey with domain experts. The detected convertibilities and produced converters were validated as relevant from a biological point of view. Furthermore the automatically produced graph of potentially compatible services exhibited a connectivity higher than with the ad'hoc approaches. Indeed, the experts discovered unknown possible connexions.
    [bibtex-key = ba2015TLDKS] [bibtex-entry]


  3. Mireille Ducassé and Peggy Cellier. Using Bids, Arguments and Preferences in Sensitive Multi-unit Assignments: A p-Equitable Process and a Course Allocation Case Study. Journal of Group Decision and Negotiation, 25(6):1211-1235, 2016. [WWW] [doi:10.1007/s10726-016-9483-9] Keyword(s): group decision support, thinklet, formal concept analysis, logical information systems.
    Abstract:
    Bonus distribution in enterprises or course allocation at universities are examples of sensitive multi-unit assignment problems, where a set of resources is to be allocated among a set of agents having multi-unit demands. Automatic processes exist, based on quantitative information, for example bids or preference ranking, or even on lotteries. In sensitive cases, however, decisions are taken by persons also using qualitative information. At present, no multi-unit assignment system supports both quantitative and qualitative information. In this paper, we propose \muaplis, an interactive process for multi-assignment problems where, in addition to bids and preferences, agents can give arguments to motivate their choices. Bids are used to automatically make pre-assignments, qualitative arguments and preferences help decision makers break ties in a founded way. A group decision support system, based on Logical Information Systems, allows decision makers to handle bids, arguments and preferences in a unified interface. We say that a process is {\em p-equitable} for a property $p$ if all agents satisfying $p$ are treated equally. We formally demonstrate that \muaplis{} is p-equitable for a number of properties on bids, arguments and preferences. It is also Pareto-efficient and Gale-Shapley-stable with respect to bids. A successful course allocation case study is reported. It spans over two university years. The decision makers were confident about the process and the resulting assignment. Furthermore, the students, even the ones who did not get all their wishes, found the process to be equitable.
    [bibtex-key = ducasse2016] [bibtex-entry]


  4. Peggy Cellier, Sébastien Ferré, Annie Foret, and Olivier Ridoux. Exploration des Données du Défi EGC 2016 à l'aide d'un Système d'Information Logique. In Cyril de Runz and Bruno Crémilleux, editors, Journées Francophones Extraction et Gestion des Connaissances, EGC, RNTI E-30, pages 443-448, 2016. Hermann-Éditions. [WWW] [bibtex-key = CFFR2016egc] [bibtex-entry]


  5. Sébastien Ferré. An RDF Design Pattern for the Structural Representation and Querying of Expressions. In Int. Conf. Knowledge Engineering and Knowledge Management, LNAI 10024, 2016. Springer. [WWW] Keyword(s): expression, knowledge representation, blank node, querying, RDF, Turtle, SPARQL, mathematical formulas.
    Abstract:
    Expressions, such as mathematical formulae, logical axioms, or structured queries, account for a large part of human knowledge. It is therefore desirable to allow for their representation and querying with Semantic Web technologies. We propose an RDF design pattern that fulfills three objectives. The first objective is the structural representation of expressions in standard RDF, so that expressive structural search is made possible. We propose simple Turtle and SPARQL abbreviations for the concise notation of such RDF expressions. The second objective is the automated generation of expression labels that are close to usual notations. The third objective is the compatibility with existing practice and legacy data in the Semantic Web (e.g., SPIN, OWL/RDF). We show the benefits for RDF tools to support this design pattern with the extension of SEWELIS, a tool for guided exploration and edition, and its application to mathematical search.
    [bibtex-key = Fer2016ekaw:Expr] [bibtex-entry]


  6. Sébastien Ferré. Bridging the Gap Between Formal Languages and Natural Languages with Zippers. In Harald Sack, Eva Blomqvist, Mathieu d'Aquin, Chiara Ghidini, Simone Paolo Ponzetto, and Christoph Lange, editors, The Semantic Web (ESWC). Latest Advances and New Domains, LNCS 9678, pages 269-284, 2016. Springer. [WWW] [doi:10.1007/978-3-319-34129-3_17] Keyword(s): semantic web, formal lamguage, natural language, zipper, Montague grammar.
    Abstract:
    The Semantic Web is founded on a number of Formal Languages (FL) whose benefits are precision, lack of ambiguity, and ability to automate reasoning tasks such as inference or query answering. This however poses the challenge of mediation between machines and users because the latter generally prefer Natural Languages (NL) for accessing and authoring knowledge. In this paper, we introduce the NF design pattern based on Abstract Syntax Trees (AST), Huet's zippers and Montague grammars to zip together a natural language and a formal language. Unlike question answering, translation does not go from NL to FL, but as symbol NF suggests, from ASTs (A) of an intermediate language to both NL (NF). ASTs are built interactively and incrementally through a user-machine dialog where the user only sees NL, and the machine only sees FL.
    [bibtex-key = Fer2016eswc] [bibtex-entry]


  7. Sébastien Ferré. SPARKLIS on QALD-6 Statistical Questions. In Semantic Web Evaluation Challenge, pages 178-187, 2016. Springer. Keyword(s): semantic web, QALD, statistical questions, OLAP.
    Abstract:
    This work focuses on the statistical questions introduced by the QALD-6 challenge. With the growing amout of semantic data, including numerical data, the need for RDF analytics beyond semantic search becomes a key issue of the Semantic Web. We have extended SPARKLIS from semantic search to RDF analytics by covering the computation features of SPARQL (expressions, aggregations and groupings). We could therefore participate to the new task on statistical questions, and we report the achieved performance of SPARKLIS. Compared to other participants, SPARKLIS does not translate spontaneous questions by users, but instead guide users in the construction of a question. Guidance is based on the actual RDF data so as to ensure that built questions are well-formed, non-ambiguous, and inhabited with answers. We show that SPARKLIS enables superior results for both an expert user (94\% correct) and a beginner user (76\% correct).
    [bibtex-key = Fer2016qald6] [bibtex-entry]


  8. Sébastien Ferré. Semantic Authoring of Ontologies by Exploration and Elimination of Possible Worlds. In Int. Conf. Knowledge Engineering and Knowledge Management, LNAI 10024, 2016. Springer. [WWW] Keyword(s): ontology authoring, semantic web, description logics, OWL, possible world explorer.
    Abstract:
    We propose a novel approach to ontology authoring that is centered on semantics rather than on syntax. Instead of writing axioms formalizing a domain, the expert is invited to explore the possible worlds of her ontology, and to eliminate those that do not conform to her knowledge. Each elimination generates an axiom that is automatically derived from the explored situation. We have implemented the approach in prototype PEW (Possible World Explorer), and conducted a user study comparing it to Protégé. The results show that more axioms are produced with PEW, without making more errors. More importantly, the produced ontologies are more complete, and hence more deductively powerful, because more negative constraints are expressed.
    [bibtex-key = Fer2016ekaw:Pew] [bibtex-entry]


  9. Sébastien Ferré and Peggy Cellier. Graph-FCA in Practice. In O. Haemmerlé, G. Stapleton, and C. Faron-Zucker, editors, Int. Conf. Conceptual Structures (ICCS) - Graph-Based Representation and Reasoning, LNCS 9717, pages 107-121, 2016. Springer. [WWW] [doi:10.1007/978-3-319-40985-6_9] Keyword(s): formal concept analysis, knowledge graph, graph pattern, algorithm.
    Abstract:
    With the rise of the Semantic Web, more and more relational data are made available in the form of knowledge graphs (e.g., RDF, conceptual graphs). A challenge is to discover conceptual structures in those graphs, in the same way as Formal Concept Analysis (FCA) discovers conceptual structures in tables. Graph-FCA has been introduced in a previous work as an extension of FCA for such knowledge graphs. In this paper, algorithmic aspects and use cases are explored in order to study the feasability and usefulness of G-FCA. We consider two use cases. The first one extracts linguistic structures from parse trees, comparing two graph models. The second one extracts workflow patterns from cooking recipes, highlighting the benefits of n-ary relationships and concepts.
    [bibtex-key = FerCel2016iccs] [bibtex-entry]


  10. Clément Gautrais, Peggy Cellier, Thomas Guyet, René Quiniou, and Alexandre Termier. Understanding Customer Attrition at an Individual Level: a New Model in Grocery Retail Context. In Proceedings of the 19th International Conference on Extending Database Technology, EDBT 2016, Bordeaux, France, March 15-16, 2016, Bordeaux, France, March 15-16, 2016., pages 686-687, 2016. [WWW] [bibtex-key = GautraisCGQT16] [bibtex-entry]


  11. Pierre Maillot, Sébastien Ferré, Peggy Cellier, Mireille Ducassé, and Franck Partouche. FORMULIS: Dynamic Form-Based Interface For Guided Knowledge Graph Authoring. In 20th International Conference on Knowledge Engineering and Knowledge Management, Posters & Demonstrations, 2016. [WWW] Keyword(s): semantic web, form, knowledge authoring, user interface.
    Abstract:
    Knowledge acquisition is a central issue of the Semantic Web. Knowledge cannot always be automatically extracted from existing data, thus contributors have to make efforts to create new data. In this paper, we propose FORMULIS, a dynamic form-based interface designed to make RDF data authoring easier. FORMULIS guides contributors through the creation of RDF data by suggesting fields and values according to the previously filled fields and the previously created resources.
    [bibtex-key = formulis16ekaw] [bibtex-entry]


2015
  1. Peggy Cellier, Thierry Charnois, Andreas Hotho, Stan Matwin, Marie-Francine Moens, and Yannick Toussaint, editors. Proceedings of the Workshop on Interactions between Data Mining and Natural Language Processing, DMNLP 2015, co-located with the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML-PKDD 2015, volume 1410 of CEUR Workshop Proceedings, 2015. CEUR-WS.org. [bibtex-key = 2015dmnlp] [bibtex-entry]


  2. Mouhamadou Ba. Composition guidée de services : application aux workflows d'analyse de données en bio-informatique. PhD thesis, Thèse de l'INSA Rennes - École doctorale MATISSE, 4 décembre 2015. Note: Supervisée par M. Ducassé et S. Ferré.
    Abstract:
    Dans les domaines scientifiques, particulièrement en bioinformatique, des services élémentaires sont composés sous forme de workflows pour effectuer des expériences d'analyse de données complexes. À cause de l'hétérogénéité des ressources, la composition de services est une tâche difficile. Les utilisateurs, en composant des workflows, manquent d'assistance pour retrouver et interconnecter les services compatibles. Les solutions existantes utilisent des services spéciaux définis de manière manuelle pour gérer les conversions de formats de données entre les entrées et sorties des services dans les workflows. Cela est pénible pour un utilisateur final. Gérer les incompatibilités des services avec des convertisseurs manuels prend du temps et est lourd. Il existe des solutions automatisées pour faciliter la composition de workflows mais elles sont généralement limitées dans le guidage et l'adaptation des données entre services. La première contribution de cette thèse propose de détecter systématiquement la convertibilité des sorties vers les entrées des services. La détection de convertibilité repose sur un système de règles basé sur une abstraction des types d'entrée et sortie des services. L'abstraction de types permet de considérer la nature et la composition des données d'entrée et sortie. Les règles permettent la décomposition et la composition ainsi que la spécialisation et la généralisation de types. Elles permettent également de générer des convertisseurs de données à utiliser entre services dans les workflows. La deuxième contribution propose une approche interactive qui permet de guider des utilisateurs à composer des workflows en fournissant des suggestions de services et de liaisons compatibles basées sur la convertibilité de types d'entrée et sortie des services. L'approche est basée sur le modèle des Systèmes d'Information Logiques (LIS) qui permettent des requêtes et une navigation guidées et sûres sur des données représentées avec une logique uniforme. Avec notre approche, la composition de workflows est sûre et complète vis-à-vis de propriétés désirées. Les résultats et les expériences, effectués sur des services et des types de données en bioinformatique, montrent la pertinence de nos approches. Nos approches offrent des mécanismes adaptés pour gérer les incompatibilités de services dans les workflows, en prenant en compte la structure composite des données d'entrée et sortie. Elles permettent également de guider, étape par étape, des utilisateurs à définir des workflows bien formés à travers des suggestions pertinentes.
    [bibtex-key = Ba2015PhD] [bibtex-entry]


  3. Peggy Cellier, Thierry Charnois, Marc Plantevit, Christophe Rigotti, Bruno Crémilleux, Olivier Gandrillon, Jirì Kléma, and Jean-Luc Manguin. Sequential pattern mining for discovering gene interactions and their contextual information from biomedical texts. J. Biomedical Semantics, 6:27, 2015. [bibtex-key = CellierCPRCGKM15] [bibtex-entry]


  4. Mouhamadou Ba, Sébastien Ferré, and Mireille Ducassé. Safe Suggestions Based on Type Convertibility to Guide Workflow Composition. In F. Esposito, O. Pivert, M.-S. Hacid, Z.W. Ras, and S. Ferilli, editors, Int. Symp. Foundations of Intelligent Systems (ISMIS), LNCS 9384, 2015. Springer.
    Abstract:
    This paper proposes an interactive approach that guides users in the step-by-step composition of services by providing safe suggestions based on type convertibility. % Users specify the points of the workflow (called the focus) they want to complete, and our approach suggests services and connections whose data types are compatible with the focus. % We prove the safeness (every step produces a well-formed workflow) and the completeness (every well-formed workflow can be built) of our approach.
    [bibtex-key = ba2015ISMIS] [bibtex-entry]


  5. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Sequence mining under multiple constraints. In Roger L. Wainwright, Juan Manuel Corchado, Alessio Bechini, and Jiman Hong, editors, ACM Symposium on Applied Computing, pages 908-914, 2015. ACM. [bibtex-key = BechetCCC15] [bibtex-entry]


  6. Soda Cissé, Olivier Ridoux, and Peggy Cellier. Représentation Symbolique de Trajectoires Géolocalisées. In Hajer Baazaoui Zghal, Christophe Claramunt, and Mohamed Ramzi Haddad, editors, Int. Conf. Spatial Analysis and GEOmatics (SAGEO), volume 1535 of CEUR Workshop Proceedings, pages 208-221, 2015. CEUR-WS.org. [bibtex-key = CisseCR15] [bibtex-entry]


  7. Sébastien Ferré. A Proposal for Extending Formal Concept Analysis to Knowledge Graphs. In J. Baixeries, C. Sacarea, and M. Ojeda-Aciego, editors, Int. Conf. Formal Concept Analysis (ICFCA), LNCS 9113, pages 271-286, 2015. Springer. Keyword(s): formal concept analysis, knowledge graph, Semantic Web, graph pattern, relation, projection.
    Abstract:
    Knowledge graphs offer a versatile knowledge representation, and have been studied under different forms, such as conceptual graphs or Datalog databases. With the rise of the Semantic Web, more and more data are available as knowledge graphs. FCA has been successful for analyzing, mining, learning, and exploring tabular data, and our aim is to help transpose those results to graph-based data. Previous FCA approaches have already addressed relational data, hence graphs, but with various limits. We propose G-FCA as an extension of FCA where the formal context is a knowledge graph based on n-ary relationships. The main contributions is the introduction of ``n-ary concepts'', i.e. concepts whose extents are n-ary relations of objects. Their intents, ``projected graph patterns'', mix relationships of different arities, objects, and variables. In this paper, we lay first theoretical results, in particular the existence of a concept lattice for each concept arity, and the role of relational projections to connect those different lattices.
    [bibtex-key = Fer2015icfca] [bibtex-entry]


  8. Sébastien Ferré. Conception interactive d'ontologies par élimination de mondes possibles. In Ingénierie des connaissances (IC), 2015. AFIA. [WWW] Keyword(s): Semantic Web, ontology, OWL, design, syntax/semantic gap, interaction.
    Abstract:
    La conception d'ontologies constitue souvent un frein à l'adoption des techniques de l'ingénierie des connaissances et du Web sémantique. Une raison est bien sûr l'emploi de formalismes et des concepts logiques qui y sont associés. Une autre raison qui nous semble plus profonde est le fossé entre syntaxe et sémantique, c'est-à-dire entre la forme de surface de l'ontologie (axiomes) et ce qu'elle rend nécessaire/possible/impossible (modèles). Ce fossé entraîne des divergences entre l'intention du concepteur et sa modélisation qui se manifestent par des inférences inattendues, voire des incohérences. Nous proposons une nouvelle approche de conception d'ontologies fondée sur l'exploration et l'élimination interactive de ``mondes possibles'' (modèles). Elle réduit le fossé syntaxe/sémantique en interdisant par construction la production d'incohérence, et en montrant en permanence au concepteur ce qui peut être inféré ou non. Un prototype, PEW (Possible World Explorer), permet d'expérimenter cette approche et de la comparer à d'autres éditeurs d'ontologies.
    [bibtex-key = Fer2015ic] [bibtex-entry]


  9. Annie Foret. A Logical Information System Proposal for Browsing Terminological Resources. In T. Poibeau and P. Faber, editors, Int. Conf. Terminology and Artificial Intelligence, volume 1495 of CEUR Workshop Proceedings, pages 51-59, 2015. CEUR-WS.org. Note: Keywords : Scientific terminology, Technological terminology, Multilingual applications, Information extraction, Textual data mining, Information retrieval, Linguistic resources, Open Data, Information Quality, Legal Information. [WWW]
    Abstract:
    This article presents an automated construction of a logical information context from a terminological resource, available in xml ; we apply this to the resource FranceTerme and to Camelis tool and we discuss how the resulting context can be used with such a tool dedicated to logical contexts. The purpose of this development and the choices related to this experiment is twofold : to facilitate the use of a rich linguistic resource available as open-data in xml ; to test and envision a systematic transformation of such xml resources to logical contexts. A logical view of a context allows to explore information in a flexible way, without writing explicit queries, it may also provide insights on the quality of the data. Such a context can be enriched by other information (of diverse natures), it can also be linked with other applications (according to arguments supplied by the context).
    [bibtex-key = DBLP:conf/tia/Foret15] [bibtex-entry]


  10. Annie Foret. A Logical Information System Proposal for Browsing Terminological Resources, Demo session. In T. Poibeau and P. Faber, editors, Int. Conf. Terminology and Artificial Intelligence, volume 1495 of CEUR Workshop Proceedings, 2015. CEUR-WS.org. [WWW] [bibtex-key = DBLP:conf/tia/Foret15d] [bibtex-entry]


  11. Annie Foret. TermLis : un contexte d'information logique pour des ressources terminologiques. In Conférence sur le Traitement Automatique des Langues Naturelles, pages 642-643, 2015. Association pour le Traitement Automatique des Langues. Note: Keywords:Multilingual applications, Classification, Information extraction, Textual data mining, Information retrieval, Linguistic resources, Open Data, Information Quality, Legal Information. [WWW]
    Abstract:
    We present TermLis a logical information context constructed from terminological resources available in XML (FranceTerme), for a flexible use with a logical context system (CAMELIS). A logical view of a context allows to explore information in a flexible way, without writing explicit queries, it may also provide insights on the quality of the data. Such a context can be enriched by other information (of diverse natures), it can also be linked with other applications (according to arguments supplied by the context). We show how to use TermLis and we illustrate, through this concrete realization from FranceTerme data, the advantages of such an approach with terminological data.
    [bibtex-key = foret:2015:TALN] [bibtex-entry]


  12. Annie Foret, Valérie Bellynck, and Christian Boitet. Akenou-Breizh, un projet de plate-forme valorisant des ressources et outils informatiques et linguistiques pour le breton. In Actes du Traitement Automatique des Langues Régionales de France et d'Europe, 2015. Association pour le Traitement Automatique des Langues. Note: Keywords : Breton, heritage language, usage language, tools and resources, contrastive studies. [WWW]
    Abstract:
    We present a new project, Akenou-Breizh, that aims to (1) put in place a platform allowing to study the influences of an heritage language, such as Breton, on a usage language, such as French, and (2) to make available, to all interested persons, tools well integrated in the "semantic and multilingual web" and proposing proactive access to various kinds of knowledge concerning Breton, as well as direct visualisation of infrasentential correspondences in aligned bilingual presentations. We plan not only to use the numerous freely available resources, in particular those of OPLB and of the APERTIUM project, but also to create new ones, such as good quality bilingual aligned corpora, thereby using the "collaborative web", and to build on the dedicated lingwarium.org web site linguistic modules improving on or extending those that exist, for example a morphological analyzer-generator.We also describe an experiment set up starting from a reduced lexicon for Breton, that shows how it is possible to enrich a classical dictionary, by linking it to a lattice of topics and to a context management system (here CAMELIS), in such a way one can query it (along semantic facets) and compare different resources.
    [bibtex-key = foret-bellynck-boitet:2015:TALaRE] [bibtex-entry]


  13. Guillaume Aucher and Olivier Ridoux. Manifest: For more Interactions between Computer Science, Human and Social Sciences and Arts. Science des données et sciences humaines et sociales : atelier 4 du Data Science Symposium, November 2015. [WWW] [PDF] Keyword(s): Art, Computer science, Human and social sciences, Interdisciplinarity research. [bibtex-key = AucherRidoux2015manifest] [bibtex-entry]


2014
  1. Peggy Cellier, Thierry Charnois, Andreas Hotho, Stan Matwin, Marie-Francine Moens, and Yannick Toussaint, editors. Proceedings of the Workshop on Interactions between Data Mining and Natural Language Processing, DMNLP 2014, co-located with the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML-PKDD 2014, volume 1202 of CEUR Workshop Proceedings, 2014. CEUR-WS.org. [bibtex-key = 2014dmnlp] [bibtex-entry]


  2. Sébastien Ferré. Reconciling Expressivity and Usability in Information Access - From Filesystems to the Semantic Web. Habilitation thesis, Matisse, Univ. Rennes 1, 2014. Note: Habilitation à Diriger des Recherches (HDR), defended on November 6th. Keyword(s): expressivity, usability, information access, information retrieval, query language, navigation structure, interactive view, abstract conceptual navigation, file system, semantic web.
    Abstract:
    In many domains where information access plays a central role, there is a gap between expert users who can ask complex questions through formal query languages (e.g., SQL), and lay users who either are dependent on expert users, or must restrict themselves to ask simpler questions (e.g., keyword search). Because of the formal nature of those languages, there seems to be an unescapable trade-off between expressivity and usability in information systems. The objective of this thesis is to present a number of results and perspectives that show that the expressivity of formal languages can be reconciled with the usability of widespread information systems (e.g., browsing, Faceted Search (FS)). The final aim of this work is to empower people with the capability to produce, explore, and analyze their data in a powerful way. \paragraph{} We have proposed a number of theories and implementations to better reconcile expressivity and usability, and applied them to a number of contexts going from file systems to the Semantic Web. In this thesis, we introduce an unifying framework inspired by Formal Concept Analysis (FCA) to factor out the main ideas of all those results: Abstract Conceptual Navigation (ACN). The principle of ACN is to guide users by letting them {\em navigate} in a conceptual space where places are {\em concepts} connected by navigation links. Concepts are characterized by a formal query, and are made of two parts: an {\em extension} and an {\em intension}. The extension is made of query results while the intension is made of the query itself and an index of query increments over results. Finally, navigation links are formally defined as query transformations. The conceptual space is not static but is induced by concrete data, and evolves with it. ACN therefore combines the {\em expressivity} of formal query languages with the {\em guidance} of conceptual navigation. The {\em readability} of queries is improved by verbalizing them to (or parsing them from) a Controlled Natural Language (CNL). Readability and guidance together support usability by speaking user's language, and by providing a systematic assistance.
    [bibtex-key = Fer2014hdr] [bibtex-entry]


  3. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Fouille de motifs séquentiels pour la découverte de relations entre gènes et maladies rares. Revue d'Intelligence Artificielle, 28(2-3):245-270, 2014. Keyword(s): data mining, sequential patterns, information extraction, linguistic patterns, rare diseases.
    Abstract:
    Orphanet est un organisme dont l'objectif est notamment de rassembler des collections d'articles traitant de maladies rares. Cependant, l'acquisition de nouvelles connaissances dans ce domaine est actuellement réalisée manuellement. Dès lors, obtenir de nouvelles informations relatives aux maladies rares est un processus chronophage. Permettre d'obtenir ces informations de manière automatique est donc un enjeu important. Dans ce contexte, nous proposons d'aborder la question de l'extraction de relations entre gènes et maladies rares en utilisant des approches de fouille de données, plus particulièrement de fouille de motifs séquentiels sous contraintes. Nos expérimentations montrent l'intérêt de notre approche pour l'extraction de relations entre gènes et maladies rares à partir de résumés d'articles de PubMed.
    [bibtex-key = BCCC14] [bibtex-entry]


  4. Mireille Ducassé and Peggy Cellier. Fair and Fast Convergence on Islands of Agreement in Multicriteria Group Decision Making by Logical Navigation. Group Decision and Negotiation, 23(4):673-694, July 2014. [WWW] [doi:10.1007/s10726-013-9372-4] Keyword(s): Multicriteria Decision, Multicriteria Sorting, Consensus Reaching, Group Decision Support System, ThinkLets, Logical Information Systems, Formal Concept Analysis.
    Abstract:
    Reasoning on multiple criteria is a key issue in group decision to take into account the multidimensional nature of real-world decision-making problems. In order to reduce the induced information overload, in multicriteria decision analysis, criteria are in general aggregated, in many cases by a simple discriminant function of the form of a weighted sum. It requires to, a priori and completely, elicit preferences of decision makers. That can be quite arbitrary. In everyday life, to reduce information overload people often use a heuristic, called ``Take-the-best'': they take criteria in a predefined order, the first criterion which discriminates the alternatives at stake is used to make the decision. Although useful, the heuristic can be biased. This article proposes the Logical Multicriteria Sort process to support multicriteria sorting within islands of agreement. It therefore does not require a complete and consistent a priori set of preferences, but rather supports groups to quickly identify the criteria for which an agreement exists. The process can be seen as a generalization of Take-the-best. It also proposes to consider one criterion at a time but once a criterion has been found discriminating it is recorded, the process is iterated and relevant criteria are logically combined. Hence, the biases of Take-the-best are reduced. The process is supported by a GDSS, based on Logical Information Systems, which gives instantaneous feedbacks of each small decision and keeps tracks of all of the decisions taken so far. The process is incremental, each step involves low information load. It guarantees some fairness because all considered alternatives are systematically analyzed along the selected criteria. A successful case study is reported.
    [bibtex-key = ducasse2013b] [bibtex-entry]


  5. Sébastien Ferré. SQUALL: The expressiveness of SPARQL 1.1 made available as a controlled natural language. Data & Knowledge Engineering, 94:163-188, 2014. [WWW] [doi:10.1016/j.datak.2014.07.010] Keyword(s): controlled natural language, semantic web, RDF, SPARQL, expressiveness.
    Abstract:
    The Semantic Web (SW) is now made of billions of triples, which are available as Linked Open Data (LOD) or as RDF stores. The SPARQL query language provides a very expressive way to search and explore this wealth of semantic data. However, user-friendly interfaces are needed to bridge the gap between end-users and SW formalisms. Navigation-based interfaces and natural language interfaces require no or little training, but they cover a small fragment of SPARQL's expressivity. We propose SQUALL, a query and update language that provides the full expressiveness of SPARQL~1.1 through a flexible controlled natural language (e.g., solution modifiers through superlatives, relational algebra through coordinations, filters through comparatives). A comprehensive and modular definition is given as a Montague grammar, and an evaluation of naturalness is done on the QALD challenge. SQUALL is conceived as a component of natural language interfaces, to be combined with lexicons, guided input, and contextual disambiguation. It is available as a Web service that translates SQUALL sentences to SPARQL, and submits them to SPARQL endpoints (e.g., DBpedia), therefore ensuring SW compliance, and leveraging the efficiency of SPARQL engines.
    [bibtex-key = Fer2014dke] [bibtex-entry]


  6. Mouhamadou Ba, Sébastien Ferré, and Mireille Ducassé. Convertibility between input and output types to help compose services in bioinformatics. In Colloque africain sur la recherche en informatique et mathématiques appliquées (CARI), pages 141-148, 2014. [bibtex-key = ba2014cari] [bibtex-entry]


  7. Mouhamadou Ba, Sébastien Ferré, and Mireille Ducassé. Convertibilité entre types d'entrée et de sortie pour la composition de services en bio-informatique. In Conf. Reconnaissance de Formes et Intelligence Artificielle (RFIA), 2014. [bibtex-key = ba2014rfia] [bibtex-entry]


  8. Mouhamadou Ba, Sébastien Ferré, and Mireille Ducassé. Generating Data Converters to Help Compose Services in Bioinformatics Workflows. In Hendrick Decker et al., editor, Int. Conf. Database and Expert Systems Applications (DEXA), LNCS 8644, pages 284-298, 2014. Springer. Keyword(s): workflow, bioinformatic, data converter, convertibility, rule system.
    Abstract:
    Heterogeneity of data and data formats in bioinformatics often entail a mismatch between inputs and outputs of different services, making it difficult to compose them into workflows. To reduce those mismatches bioinformatics platforms propose ad'hoc converters written by hand. This article proposes to systematically detect convertibility from output types to input types. Convertibility detection relies on abstract types, close to XML Schema, allowing to abstract data while precisely accounting for its composite structure. Detection is accompanied by an automatic generation of converters between input and output XML data. Our experiment on bioinformatics services and datatypes, performed with an implementation of our approach, shows that the detected convertibilities and produced converters are relevant from a biological point of view. Furthermore they automatically produce a graph of potentially compatible services with a connectivity higher than with the ad'hoc approaches.
    [bibtex-key = ba2014dexa] [bibtex-entry]


  9. Soda Cissé, Peggy Cellier, and Olivier Ridoux. Segmentation of Geolocalized Trajectories using Exponential Moving Average. In Colloque Africain sur la Recherche en Informatique et Mathématiques Appliquées (CARI), pages 149-156, 2014. Keyword(s): geolocalized trajectories, segmentation.
    Abstract:
    Nowadays, large sets of data describing trajectories of mobile objects are made available by the generalization of geolocalisation sensors. Relevant information, for instance, the most used routes by children to go to school or the most extensively used streets in the morning by workers, can be extracted from this amount of available data allowing, for example, to reconsider the urban space. A trajectory is represented by a set of points (x; y; t) where x and y are the geographic coordinates of a mobile object and t is a date. These data are difficult to explore and interpret in their raw form, i.e. in the form of points (x; y; t), because they are noisy, irregularly sampled and too low level. A first step to make them usable is to resample the data, smooth it, and then to segment it into higher level segments (e.g. "stops" and "moves") that give a better grip for interpretation than the raw coordinates. In this paper, we propose a method for the segmentation of these trajectories in accelerate/decelerate segments which is based on the computation of exponential moving averages (EMA). We have conducted experiments where the exponential moving average proves to be an efficient smoothing function, and the difference between two EMA of different weights proves to discover significant accelerating-decelerating segments.
    [bibtex-key = CisseCR2014] [bibtex-entry]


  10. Mireille Ducassé and Peggy Cellier. Using Biddings and Motivations in Multi-unit Assignments. In Pascale Zaraté, Gregory E. Kersten, and Jorge E. Hernandez, editors, Group Decision and Negotiation. A Process-Oriented View, volume 180 of Lecture Notes in Business Information Processing, pages 53-61, 2014. Springer. [WWW] [doi:10.1007/978-3-319-07179-4-6] Keyword(s): group decision support, thinkLet, formal concept analysis, logical information systems, course allocation, multi-unit assignment.
    Abstract:
    In this paper, we propose a process for small to medium scale multi-assignment problems. In addition to biddings, agents can give motivations to explain their choices in order to help decision makers break ties in a founded way. A group decision support system, based on Logical Information Systems, allows decision makers to easily face both biddings and motivations. Furthermore, it guaranties that all the agents are treated equally. A successful case study about a small course assignment problem at a technical university is reported.
    [bibtex-key = ducasse2014] [bibtex-entry]


  11. Sébastien Ferré. Expressive and Scalable Query-Based Faceted Search over SPARQL Endpoints. In P. Mika, T. Tudorache, A. Bernstein, C. Welty, C. A. Knoblock, D. Vrandecic, P. T. Groth, N. F. Noy, K. Janowicz, and C. A. Goble, editors, The Semantic Web (ISWC), LNCS 8797, pages 438-453, 2014. Springer. Note: Nominee for the best research paper award. [WWW] Keyword(s): SPARQL endpoints, semantic search, faceted search, user interaction, SPARQL queries, query-based faceted search, expressivity, scalability, portability, usability, Sparklis.
    Abstract:
    Linked data is increasingly available through SPARQL endpoints, but exploration and question answering by regular Web users largely remain an open challenge. Users have to choose between the expressivity of formal languages such as SPARQL, and the usability of tools based on navigation and visualization. In a previous work, we have proposed Query-based Faceted Search (QFS) as a way to reconcile the expressivity of formal languages and the usability of faceted search. In this paper, we further reconcile QFS with scalability and portability by building QFS over SPARQL endpoints. We also improve expressivity and readability. Many SPARQL features are now covered: multidimensional queries, union, negation, optional, filters, aggregations, ordering. Queries are now verbalized in English, so that no knowledge of SPARQL is ever necessary. All of this is implemented in a portable Web application, Sparklis, and has been evaluated on many endpoints and questions.
    [bibtex-key = Fer2014iswc] [bibtex-entry]


  12. Sébastien Ferré. SPARKLIS: a SPARQL Endpoint Explorer for Expressive Question Answering. In M. Horridge, M. Rospocher, and J. van Ossenbruggen, editors, ISWC Posters & Demonstrations Track, volume 1272 of CEUR Workshop Proceedings, pages 45-48, 2014. CEUR-WS.org. [WWW] Keyword(s): demo, SPARQL endpoints, semantic search, faceted search, user interaction, SPARQL queries, query-based faceted search, expressivity, scalability, portability, usability, Sparklis.
    Abstract:
    SPARKLIS is a Semantic Web tool that helps users explore SPARQL endpoints by guiding them in the interactive building of questions and answers, from simple ones to complex ones. It combines the fine-grained guidance of faceted search, most of the expressivity of SPARQL, and the readability of (controlled) natural languages. No endpoint-specific configuration is necessary, and no knowledge of SPARQL and the data schema is required from users. This demonstration paper is a companion to the research paper~\cite{Fer2014iswc}.
    [bibtex-key = Fer2014demo] [bibtex-entry]


  13. Annie Foret. On Associative Lambek Calculus Extended with Basic Proper Axioms. In C. Casadio, B. Coecke, M. Moortgat, and P. Scott, editors, Categories and Types in Logic, Language, and Physics - Essays Dedicated to Jim Lambek on the Occasion of His 90th Birthday, LNCS 8222, pages 172-187, 2014. Springer. [WWW] Keyword(s): lambek calculus, associativity.
    Abstract:
    The purpose of this article is to show that the associative Lambek calculus extended with basic proper axioms can be simulated by the usual associative Lambek calculus, with the same number of types per word in a grammar. An analogue result had been shown earlier for pregroups grammars (2007). We consider Lambek calculus with product, as well as the product-free version.
    [bibtex-key = Foret2014birthday] [bibtex-entry]


  14. Annie Foret. On Harmonic CCG and Pregroup Grammars. In N. Asher and S. Soloviev, editors, Int. Conf. Logical Aspects of Computational Linguistics (LACL), LNCS 8535, pages 83-95, 2014. Springer. [WWW] Keyword(s): pregroup grammar, CCG.
    Abstract:
    This paper studies mappings between CCG and pregroup grammars, to allow a transfer of linguistic resources from one formalism to the other. We focus on mappings that preserve the binary structures, we also discuss some possible alternatives in the underlying formalisms, with some experiments.
    [bibtex-key = Foret2014lacl] [bibtex-entry]


  15. Solen Quiniou, Peggy Cellier, and Thierry Charnois. Fouille de données pour associer des noms de sessions aux articles scientifiques. In Brigitte Bigi, editor, Défi Fouille de Textes - DEFT 2014 (Atelier TALN), 2014. Laboratoire Parole et Langage. Note: ISBN: 978-2-9518233-6-5. Keyword(s): data mining, sequence mining, graph mining, paper categorisation.
    Abstract:
    Nous décrivons dans cet article notre participation à l'édition 2014 de DEFT. Nous nous intéressons à la tâche consistant à associer des noms de session aux articles d'une conférence. Pour ce faire, nous proposons une approche originale, symbolique et non supervisée, de découverte de connaissances. L'approche combine des méthodes de fouille de données séquentielles et de fouille de graphes. La fouille de séquences permet d'extraire des motifs fréquents dans le but de construire des descriptions des articles et des sessions. Ces descriptions sont ensuite représentées par un graphe. Une technique de fouille de graphes appliquée sur ce graphe permet d'obtenir des collections de sous-graphes homogènes, correspondant à des collections d'articles et de noms de sessions.
    [bibtex-key = QCC14] [bibtex-entry]


2013
  1. Peggy Cellier, Felix Distel, and Bernhard Ganter, editors. Formal Concept Analysis, 11th International Conference, ICFCA 2013, Dresden, Germany, May 21-24, 2013. Proceedings, volume 7880 of Lecture Notes in Computer Science, 2013. Springer. [bibtex-key = DBLP:conf/icfca/2013] [bibtex-entry]


  2. Annie Foret. CATEGORIES AND TYPES IN LOGIC, LANGUAGE AND PHYSICS - Festschrift on the occasion of Jim Lambek's 90th birthday, chapter On Associative Lambek Calculus extended with basic proper axioms. Springer, Lecture Notes in Computer Science, New York, Casadio C., Coecke B., Moortgat M., Scott P. (editors), in print edition, 2013.
    Abstract:
    The purpose of this article is to show that the associative Lambek calculus extended with basic proper axioms can be simulated by the usual associative Lambek calculus, with the same number of types per word in a grammar. An analogue result had been shown earlier for pregroups grammars (2007). We consider Lambek calculus with product, as well as the product-free version.
    [bibtex-key = Lambek90th] [bibtex-entry]


  3. Mouhamadou Ba. Guided Composition of Tasks with Logical Information Systems - Application to Data Analysis Workflows in Bioinformatics. In P. Cimiano, O. Corcho, V. Presutti, L. Hollink, and S. Rudolph, editors, Extended Semantic Web Conf., LNCS 7882, pages 661-665, 2013. Springer. [bibtex-key = Ba2013] [bibtex-entry]


  4. Mireille Ducassé. Helping Facilitators Build on Experience When Preparing Meetings With Logical Information Systems. In Bilyana Martinovski, editor, Group Decision and Negotiation Conference, pages 139-143, 2013. Department of Computer and Systems Sciences, Stockholm University. Note: Extended abstract. Keyword(s): Artificial Intelligence Methods in GDN, Logical information systems, Concept Analysis Application, Facilitator support, Agenda building, ThinkLets.
    Abstract:
    This paper reports work in progress about using Logical Information Systems to help facilitators build on experience when preparing meetings. Features of meetings similar to the one under construction are automatically suggested without having to ask for suggestions. Suggestions take into account the whole information about all the meetings already recorded in the system as well as facilitation knowledge, such as thinkLets. Usual techniques and processes that facilitators like to use are naturally suggested. An unusual technique is suggested for example if the facilitator enters a keyword that is a feature of that technique. Although a lot remains to be done, the proposed approach already shows contributions that make believe that it is worth investigating further. The main one is that it builds on the facilitator very practice. Other important features are flexibility and adaptability.
    [bibtex-key = ducasse2013] [bibtex-entry]


  5. Sébastien Ferré. Representation of Complex Expressions in RDF. In P. Cimiano, M. Fernández, V. Lopez, S. Schlobach, and J. Völker, editors, Extended Semantic Web Conf. (ESWC Satellite Events), LNCS 7955, pages 273-274, 2013. Springer. [PDF] Keyword(s): semantic web, RDF, blank nodes, expressions, knowledge representation, querying, query-based faceted search, mathematical search. [bibtex-key = Fer2013eswc] [bibtex-entry]


  6. Sébastien Ferré. SQUALL: A Controlled Natural Language as Expressive as SPARQL 1.1. In E. Métais, F. Meziane, M. Saraee, V. Sugumaran, and S. Vadera, editors, Int. Conf. Applications of Natural Language to Information System (NLDB), LNCS 7934, pages 114-125, 2013. Springer. [PDF] Keyword(s): controlled natural language, query language, update language, semantic web, SPARQL, expressivity.
    Abstract:
    The Semantic Web is now made of billions of triples, which are available as Linked Open Data (LOD) or as RDF stores. The most common approach to access RDF datasets is through SPARQL, an expressive query language. However, SPARQL is difficult to learn for most users because it exhibits low-level notions of relational algebra such as union, filters, or grouping. We present SQUALL, a high-level language for querying and updating an RDF dataset. It has a strong compliance with RDF, covers all features of SPARQL 1.1, and has a controlled natural language syntax that completely abstracts from low-level notions. SQUALL is available as two web services: one for translating a SQUALL sentence to a SPARQL query or update, and another for directly querying a SPARQL endpoint such as DBpedia.
    [bibtex-key = Fer2013nldb] [bibtex-entry]


  7. Alice Hermann, Mireille Ducassé, Sébastien Ferré, and Jean Lieber. Une approche fondée sur le raisonnement à partir de cas pour la mise à jour interactive d'objets du Web sémantique. In 21ème atelier Français de Raisonnement à Partir de Cas (RàPC), Lille, France, 2013. [WWW] Keyword(s): mise à jour d'objets, Web sémantique, RDFS, formulaire de saisie, recherche de cas par relâchement. [bibtex-key = HerDucFerLie2013] [bibtex-entry]


  8. Solen Quiniou, Peggy Cellier, Thierry Charnois, and Dominique Legallois. Graph Mining under Linguistic Constraints to Explore Large Texts. In Intelligent Text Processing and Computational Linguistics (CICLing'13), 2013. [bibtex-key = QuiniouCCL13] [bibtex-entry]


  9. Joris Guyonvarch, Sébastien Ferré, and Mireille Ducassé. Scalable Query-based Faceted Search on top of SPARQL Endpoints for Guided and Expressive Semantic Search. Research report PI-2009, LIS - IRISA, October 2013. [WWW] [PDF] Keyword(s): Web of data, semantic search, querying, faceted search, SPARQL, SEWELIS, expressivity, usability, scalability.
    Abstract:
    Because the Web of Documents is composed of structured pages that are not meaningful to machines, search in the Web of Documents is generally processed by keywords. However, because the Web of Data provides structured information, search in the Web of Data can be more precise. SPARQL is the standard query language for querying this structured information. SPARQL is expressive and its syntax is similar to SQL. However, casual user can not write SPARQL queries. Sewelis is a search system for the Web of Data offering to explore data progressively and more user-friendly than SPARQL. Sewelis guides the search with a query built incrementally because users only have to select query elements in order to complete the query. However, Sewelis does not scale to large datasets such as DBpedia, which is composed of about 2 billion triples. In this report, we introduce Scalewelis. Scalewelis is a search system for the Web of Data that is similar to Sewelis but scalable. Moreover, Scalewelis is independent to data because it connects to SPARQL endpoints. We took part in a challenge on DBpedia with Scalewelis. We were able to answer to 70 questions out of 99 with acceptable response times.
    [bibtex-key = GuyFerDuc2013pi] [bibtex-entry]


  10. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Extraction de motifs séquentiels sous contraintes multiples, 2013. Keyword(s): data mining, sequential patterns, linguistic pattern, natural language processing. [bibtex-key = BCCC13_1] [bibtex-entry]


  11. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. SDMC : un outil en ligne d extraction de motifs se quentiels pour la fouille de textes, 2013. Keyword(s): data mining, sequential patterns, linguistic pattern, natural language processing. [bibtex-key = BCCC13_2] [bibtex-entry]


  12. Sébastien Ferré. squall2sparql: a Translator from Controlled English to Full SPARQL 1.1. Work. Multilingual Question Answering over Linked Data (QALD-3), 2013. Note: See Online Working Notes at www.clef2013.org. [PDF] Keyword(s): squall, controlled natural language, question answering, SPARQL.
    Abstract:
    This paper reports on the participation of the system {\sc squall2sparql} in the QALD-3 question answering challenge for DBpedia. {\sc squall2sparql} is a translator from SQUALL, a controlled natural language for English, to SPARQL 1.1, a standard expressive query and update language for linked open data. It covers nearly all features of SPARQL 1.1, and is directly applicable to any SPARQL endpoint.
    [bibtex-key = QALD3:SQUALL] [bibtex-entry]


  13. Joris Guyonvarc'h and Sébastien Ferré. Scalewelis: a Scalable Query-based Faceted Search System on Top of SPARQL Endpoints. Work. Multilingual Question Answering over Linked Data (QALD-3), 2013. Note: See Online Working Notes at www.clef2013.org. [PDF] Keyword(s): faceted search, question answering, SEWELIS, SPARQL.
    Abstract:
    This paper overviews the participation of Scalewelis in the QALD-3 open challenge. Scalewelis is a Faceted Search system. Faceted Search systems refine the result set at each navigation step. In Scalewelis, refinements are syntactic operations that modify the user query. Scalewelis uses the Semantic Web standards (URI, RDF, SPARQL) and connects to SPARQL endpoints.
    [bibtex-key = QALD3:SCALEWELIS] [bibtex-entry]


2012
  1. Annie Foret. On some classes of type-logical grammars that model syntax. Habilitation thesis, Matisse, Univ. Rennes 1, 2012. Note: Habilitation à Diriger des Recherches (HDR), defended on July 5th. [bibtex-key = Foret2012hdr] [bibtex-entry]


  2. Alice Hermann. Création et mise à jour d'objets dans une base de connaissances. PhD thesis, Thèse de l'INSA Rennes - École doctorale MATISSE, 17 décembre 2012. Note: Supervised by M. Ducassé and S. Ferré. [bibtex-key = Hermann2012PhD] [bibtex-entry]


  3. Olivier Bedel, Sébastien Ferré, and Olivier Ridoux. Développements logiciels en géomatique -- innovations et mutualisations, chapter GEOLIS : un système d'information logique pour l'organisation et la recherche de données géolocalisées, pages 149-180. Information géographique et Aménagement du Territoire. Hermes/Lavoisier, 2012. Keyword(s): logical information systems, geographical information systems, geo-located data. [bibtex-key = BedFerRid2012fr] [bibtex-entry]


  4. Olivier Bedel, Sébastien Ferré, and Olivier Ridoux. GEOLIS: a Logical Information System to Organize and Search Geo-Located Data. In B. Bucher and F. Le Ber, editors, Innovative Software Development in GIS, Geographical Information Systems Series, pages 151-188. Wiley, 2012. Keyword(s): logical information systems, geographical information systems, geo-located data. [bibtex-key = BedFerRid2012en] [bibtex-entry]


  5. Denis Béchet, Alexandre Dikovsky, and Annie Foret. Categorial grammars with iterated types form a strict hierarchy of k-valued languages. Theor. Comput. Sci., 450:22-30, 2012. [bibtex-key = DBLP:journals/tcs/BechetDF12] [bibtex-entry]


  6. Sébastien Ferré and Alice Hermann. Reconciling faceted search and query languages for the Semantic Web. Int. J. Metadata, Semantics and Ontologies, 7(1):37-54, 2012. Keyword(s): semantic web, query languages, faceted search, query-based faceted search.
    Abstract:
    Faceted search and querying are two well-known paradigms to search the Semantic Web. Querying languages, such as SPARQL, offer expressive means for searching RDF datasets, but they are difficult to use. Query assistants help users to write well-formed queries, but they do not prevent empty results. Faceted search supports exploratory search, i.e., guided navigation that returns rich feedbacks to users, and prevents them to fall in dead-ends (empty results). However, faceted search systems do not offer the same expressiveness as query languages. We introduce {\em Query-based Faceted Search} (QFS), the combination of an expressive query language and faceted search, to reconcile the two paradigms. We formalize the navigation of faceted search as a navigation graph, where navigation places are queries, and navigation links are query transformations. We prove that this navigation graph is {\em safe} (no dead-end), and {\em complete} (every query that is not a dead-end can be reached by navigation). In this paper, the LISQL query language generalizes existing semantic faceted search systems, and covers most features of SPARQL. A prototype, Sewelis, has been implemented, and a usability evaluation demonstrated that QFS retains the ease-of-use of faceted search, and enables users to build complex queries with little training.
    [bibtex-key = FerHer2012ijmso] [bibtex-entry]


  7. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Discovering Linguistic Patterns Using Sequence Mining. In Alexander F. Gelbukh, editor, Int. Conf. on Computational Linguistics and Intelligent Text Processing (CICLing), volume 7181 of LNCS, pages 154-165, 2012. Springer. Keyword(s): data mining, sequential patterns, linguistic pattern, natural language processing. [bibtex-key = BCCC12] [bibtex-entry]


  8. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Fouille de motifs séquentiels pour la découverte de relations entre gènes et maladies rares. In S. Szulman and J. Charlet, editors, Journées francophones d'ingénierie des connaissances, pages 149-164, 2012. INSERM UMPC. [WWW] Keyword(s): data mining, sequential patterns, relationships, bioinformatics, rare disease, natural language processing. [bibtex-key = BCCC2012ic] [bibtex-entry]


  9. Nicolas Béchet, Peggy Cellier, Thierry Charnois, and Bruno Crémilleux. Sequential Pattern Mining to Discover Relations between Genes and Rare Diseases. In IEEE Int. Symp. on Computer-Based Medical Systems (CBMS), pages 1-6, 2012. Keyword(s): data mining, sequential patterns, relationships, bioinformatics, rare disease, natural language processing. [bibtex-key = BCCC2012cbms] [bibtex-entry]


  10. Mireille Ducassé and Peggy Cellier. The LogicalMulticriteriaSort ThinkLet: Logical Navigation for Fair and Fast Convergence in Multicriteria Group Decision Making. In Adiel Teixeira de Almeida, Danielle Costa Morais, and Suzanna de França Dantas Daher, editors, Proceedings of the Group Decision and Negotiation Conference, pages 87-96, May 2012. Note: ISBN 978-85-415-0036-4. Keyword(s): multicriteria decision, logical information systems, formal concept analysis.
    Abstract:
    Information overload is a key issue in group decision. A heuristics, called ``take-the-best'', has been shown useful to face multicriteria decisions while reducing information overload: when making decisions people often take criteria in a predefined order, the first criterion which discriminates the alternatives at stake is used to make the decision. In order to rationalize group work, Briggs and de Vreede have proposed collaboration design patterns, called thinkLets. This article presents the LogicalMulticriteriaSort which can be seen as a generalization of the take-the-best heuristics. It also proposes to consider criteria one at the time but once a criterion has been found discriminating it is kept in a record, and the process is iterated. The thinkLet is supported by a GDSS, based on Logical Information Systems, which gives an instantaneous feedback of each micro decision and keeps tracks of all of the decisions taken so far. The LogicalMulticriteriaSort ThinkLet guarantees more fairness and speed than the ChauffeurSort thinkLet. It also avoids the need to give artificial values and weights to the criteria as opposed to the Multicriteria thinkLet. A successful test case is reported.
    [bibtex-key = ducasse2012] [bibtex-entry]


  11. Sébastien Ferré. Extension du langage de requêtes LISQL pour la représentation et l'exploration d'expressions mathématiques en RDF. In S. Szulman and J. Charlet, editors, Journées francophones d'ingénierie des connaissances, pages 285-300, 2012. INSERM UMPC. [WWW] Keyword(s): LISQL, query language, mathematics, RDF, semantic web, knowledge representation.
    Abstract:
    Les expressions mathématiques comptent pour une part importante dans les connaissances humaines. Nous en proposons une représentation en RDF afin de pouvoir les intégrer aux autres connaissances dans le Web sémantique. Nous étendons ensuite le langage de description et d'interrogation LISQL afin de concilier des représentations non-ambiguës, des requêtes expressives et des notations naturelles et concises. Par exemple, la requête exttt{int(...?X $\hat{~}$ 2...,?X)} permet de trouver les intégrales en~$x$ dont le corps contient la sous-expression~$x^2$. Tout cela permet d'utiliser Sewelis, un système d'information logique pour le Web sémantique, pour la représentation et l'exploration guidée d'expressions mathématiques. Ce guidage dispense les utilisateurs de maîtriser la syntaxe de LISQL et le vocabulaire tout en leur garantissant des expressions bien formées et des résultats à leurs requêtes.
    [bibtex-key = Fer2012ic] [bibtex-entry]


  12. Sébastien Ferré. SQUALL: a Controlled Natural Language for Querying and Updating RDF Graphs. In T. Kuhn and N.E. Fuchs, editors, Controlled Natural Languages, LNCS 7427, pages 11-25, 2012. Springer. Keyword(s): query language, update language, Semantic Web, RDF graphs, SPARQL, Montague grammars.
    Abstract:
    Formal languages play a central role in the Semantic Web. An important aspect regarding their design is syntax as it plays a crucial role in the wide acceptance of the Semantic Web approach. The main advantage of controlled natural languages (CNL) is to reconcile the high-level and natural syntax of natural languages, and the precision and lack of ambiguity of formal languages. In the context of the Semantic Web and Linked Open Data, CNL could not only allow more people to contribute by abstracting from the low-level details, but also make experienced people more productive, and make the produced documents easier to share and maintain. We introduce SQUALL, a controlled natural language for querying and updating RDF graphs. It has a strong adequacy with RDF, an expressiveness close to SPARQL 1.1, and a CNL syntax that completely abstracts from low-level notions such as bindings and relational algebra. We formally define the syntax and semantics of SQUALL as a Montague grammar, and its translation to SPARQL. It features disjunction, negation, quantifiers, built-in predicates, aggregations with grouping, and n-ary relations through reification.
    [bibtex-key = Fer2012cnl] [bibtex-entry]


  13. Sébastien Ferré, Pierre Allard, and Olivier Ridoux. Cubes of Concepts: Multi-dimensional Exploration of Multi-valued Contexts. In F. Domenach, D. I. Ignatov, and J. Poelmans, editors, Int. Conf. Formal Concept Analysis, LNCS 7278, pages 112-127, 2012. Springer. Keyword(s): formal concept analysis, OLAP, cubes of concepts, multi-valued contexts.
    Abstract:
    A number of information systems offer a limited exploration in that users can only navigate from one object to another object, e.g. navigating from folder to folder in file systems, or from page to page on the Web. An advantage of conceptual information systems is to provide navigation from concept to concept, and therefore from set of objects to set of objects. The main contribution of this paper is to push the exploration capability one step further, by providing navigation from set of concepts to set of concepts. Those sets of concepts are structured along a number of dimensions, thus forming a cube of concepts. We describe a number of representations of concepts, such as sets of objects, multisets of values, and aggregated values. We apply our approach to multi-valued contexts, which stand at an intermediate position between many-valued contexts and logical contexts. We explain how users can navigate from one cube of concepts to another. We show that this navigation includes and extends both conceptual navigation and OLAP operations on cubes.
    [bibtex-key = FerAllRid2012icfca] [bibtex-entry]


  14. Sébastien Ferré and Sebastian Rudolph. Advocatus Diaboli - Exploratory Enrichment of Ontologies with Negative Constraints. In A. ten Teije et al., editor, Int. Conf. Knowledge Engineering and Knowledge Management (EKAW), LNAI 7603, pages 42-56, 2012. Springer. Keyword(s): semantic web, ontology, exploratory search, negative constraints, possible world.
    Abstract:
    With the persistent deployment of ontological specifications in practice and the increasing size of the deployed ontologies, methodologies for ontology engineering are becoming more and more important. In particular, the specification of negative constraints is often neglected by the human expert, whereas they are crucial for increasing an ontology's deductive potential. % We propose a novel, arguably cognitively advantageous methodology for identifying and adding missing negative constraints to an existing ontology. To this end, a domain expert navigates through the space of satisfiable class expressions with the aim of finding absurd ones, which then can be forbidden by adding a respective constraint to the ontology. % We give the formal foundations of our approach, provide an implementation, called Possible World Explorer (PEW) and illustrate its usability by describing prototypical navigation paths using the example of the well-known pizza ontology.
    [bibtex-key = FerRud2012ekaw] [bibtex-entry]


  15. Alice Hermann, Sébastien Ferré, and Mireille Ducassé. Aide à la création d'objets dans une base RDF(S) avec des règles de relaxation. In S. Szulman and J. Charlet, editors, Journées francophones d'ingénierie des connaissances, pages 301-316, 2012. INSERM UMPC. [WWW] Keyword(s): RDF, relaxation rules, object creation, user interaction.
    Abstract:
    Quand un utilisateur crée un nouvel objet dans le Web s\'emantique, les outils existants n'exploitent ni les objets existants et leurs propri\'et\'es, ni les propri\'et\'es d\'ej\`a connues du nouvel objet. Nous proposons UTILIS, une m\'ethode d'aide \`a la cr\'eation de nouveaux objets. UTILIS cherche des objets similaires au nouvel objet en appliquant des r\`egles de relaxation \`a sa description. % Les propri\'et\'es des objets similaires servent de suggestions pour compl\'eter la description du nouvel objet. % Une \'etude utilisateur men\'ee avec des \'etudiants en master montre que les suggestions d'UTILIS ont \'et\'e utilis\'ees. Les utilisateurs ont trouv\'e les suggestions pertinentes : dans la plupart des cas, ils pouvaient trouver l'\'el\'ement recherch\'e dans les trois premiers ensembles de suggestions. De plus, ils les ont appr\'eci\'ees, car la majorit\'e souhaitent les avoir dans un \'editeur de donn\'ees du Web s\'emantique.
    [bibtex-key = HerFerDuc2012ic] [bibtex-entry]


  16. Alice Hermann, Sébastien Ferré, and Mireille Ducassé. An Interactive Guidance Process Supporting Consistent Updates of RDFS Graphs. In A. ten Teije et al., editor, Int. Conf. Knowledge Engineering and Knowledge Management (EKAW), LNAI 7603, pages 185-199, 2012. Springer. Keyword(s): semantic web, RDF, update, interactive guidance.
    Abstract:
    With existing tools, when creating a new object in the Semantic Web, users benefit neither from existing objects and their properties, nor from the already known properties of the new object. % We propose UTILIS, an interactive process to help users add new objects. While creating a new object, relaxation rules are applied to its current description to find similar objects, whose properties serve as suggestions to expand the description. % A user study conducted on a group of master students shows that students, even the ones disconcerted by the unconventional interface, used UTILIS suggestions. In most cases, they could find the searched element in the first three sets of properties of similar objects. % Moreover, with UTILIS users did not create any duplicate whereas with the other tool used in the study more than half of them did.
    [bibtex-key = HerFerDuc2012ekaw] [bibtex-entry]


  17. Alice Hermann, Sébastien Ferré, and Mireille Ducassé. Guided Semantic Annotation of Comic Panels with Sewelis. In A. ten Teije et al., editor, Knowledge Engineering and Knowledge Management (EKAW), LNCS 7603, pages 430-433, 2012. Springer. Keyword(s): demo, semantic web, RDF, update, interactive guidance.
    Abstract:
    UTILIS (Updating Through Interaction in Logical Information Systems), introduced in a research paper at EKAW'12, is an interactive process to help users create new objects in a RDF graph. While creating a new object, relaxation rules are applied to its current description to find similar objects, whose properties serve as suggestions to expand the description. UTILIS is implemented in Sewelis, a system that reconciles the expressiveness of querying languages (e.g., SPARQL), and the benefits of exploratory search found in faceted search. The same interaction principles are used for both exploration and creation of semantic data. We illustrate the UTILIS approach by applying Sewelis to the semantic annotation of comic panels, reusing the dataset that was used for a user evaluation.
    [bibtex-key = HerFerDuc2012demo] [bibtex-entry]


  18. Solen Quiniou, Peggy Cellier, Thierry Charnois, and Dominique Legallois. Fouille de données pour la stylistique : cas des motifs séquentiels émergents. In Actes des Journées Internationales d'Analyse Statistique des Données Textuelles, Liège, Belgique, pages 821-833, 2012. [WWW] Keyword(s): data mining, emerging sequential patterns, stylistic, linguistic pattern, natural language processing.
    Abstract:
    Dans cet article, nous pr{\'e}sentons une {\'e}tude sur l'utilisation de m{\'e}thodes de fouille de donn{\'e}es pour l'analyse stylistique - d'un point de vue linguistique - en consid{\'e}rant des motifs s{\'e}quentiels {\'e}mergents. Nous montrons tout d'abord que la fouille de motifs s{\'e}quentiels de mots en utilisant la contrainte gap permet d'obtenir de nouveaux patrons linguistiques pertinents par rapport aux patrons construits {\`a} partir de n-grammes. Nous {\'e}tudions ensuite l'utilisation de motifs s{\'e}quentiels d'itemsets pour produire des patrons linguistiques plus g{\'e}n{\'e}raux. Nous validons notre approche d'un point de vue quantitatif et d'un point de vue linguistique, en r{\'e}alisant des exp{\'e}rimentations sur trois corpus fran{\c c}ais correspondant {\`a} diff{\'e}rents genres de texte (la po{\'e}sie, les correspondances et les romans, respectivement). En consid{\'e}rant plus particuli{\`e}rement les textes po{\'e}tiques, nous montrons que les techniques de fouille de donn{\'e}es employ{\'e}es permettent d'identifier des patrons linguistiques caract{\'e}ristiques.
    [bibtex-key = QCCL2012JADT] [bibtex-entry]


  19. Solen Quiniou, Peggy Cellier, Thierry Charnois, and Dominique Legallois. Fouille de graphes sous contraintes linguistiques pour l'exploration de grands textes. In Actes de la Conférence sur le Traitement Automatique des Langues Naturelles, Grenoble, France, pages 253-266, 2012. [WWW] Keyword(s): Graph Mining, sentence networks, textual analysis, textual navigation, natural language processing.
    Abstract:
    Dans cet article, nous proposons une approche pour explorer des textes de taille importante en mettant en {\'e}vidence des sous-parties coh{\'e}rentes. Cette m{\'e}thode d'exploration s'appuie sur une repr{\'e}sentation en graphe du texte, en utilisant le mod{\`e}le linguistique de Hoey pour s{\'e}lectionner et apparier les phrases dans le graphe. Notre contribution porte sur l'utilisation de techniques de fouille de graphes sous contraintes pour extraire des sous-parties pertinentes du texte (c'est-{\`a}-dire des collections de sous-r{\'e}seaux phrastiques homog{\`e}nes). Nous avons r{\'e}alis{\'e} des exp{\'e}rimentations sur deux textes anglais de taille cons{\'e}quente pour montrer l'int{\'e}r{\^e}t de l'approche que nous proposons.
    [bibtex-key = QCCLTALN12] [bibtex-entry]


  20. Solen Quiniou, Peggy Cellier, Thierry Charnois, and Dominique Legallois. What About Sequential Data Mining Techniques to Identify Linguistic Patterns for Stylistics?. In Alexander F. Gelbukh, editor, Int. Conf. on Computational Linguistics and Intelligent Text Processing (CICLing), volume 7181 of LNCS, New Delhi, Inde, pages 166-177, March 2012. Springer. Keyword(s): data mining, sequential patterns, stylistic, linguistic pattern, natural language processing.
    Abstract:
    In this paper, we study the use of data mining techniques for stylistic analysis, from a linguistic point of view, by considering emerging sequential patterns. First, we show that mining sequential patterns of words with gap constraints gives new relevant linguistic patterns with respect to patterns built on n-grams. Then, we investigate how sequential patterns of itemsets can provide more generic linguistic patterns. We validate our approach from a linguistic point of view by conducting experiments on three corpora of various types of French texts (Poetry, Letters, and Fictions). By considering more particularly poetic texts, we show that characteristic linguistic patterns can be identified using data mining techniques. We also discuss how to improve our proposed approach so that it can be used more efficiently for linguistic analyses.
    [bibtex-key = QCCL12] [bibtex-entry]


  21. Annie Foret and Sébastien Ferré. On Categorial Grammars and Logical Information Systems : using CAMELIS with linguistic data. Note: Presented at the demo session of LACL'12, 2012. Keyword(s): demo.
    Abstract:
    We have explored in different perspectives on how categorial grammars can be considered as Logical Information Systems (LIS), where objects are organized and queried by logical properties, both theoretically and practically. LIS have also been considered for the development of pregroup grammars. We propose to illustrate these points with the CAMELIS tool that is an implementation of Logical Information Systems (LIS) and that has been developped at Irisa Rennes. CAMELIS may give another view on linguistic data, and provide an easy help to browse, to update, to create and to maintain or to test such data.
    [bibtex-key = demo:camelis:lacl2012] [bibtex-entry]


2011
  1. Pierre Allard. Logical modeling of multidimensional analysis of multivalued relations - Application to geographic data exploration. PhD thesis, Thèse de l'Université de Rennes 1 - École doctorale MATISSE, 12 décembre 2011. Note: Supervised by S. Ferré and O. Ridoux.
    Abstract:
    Since the beginning of data processing, the companies have realized the importance of information management solutions. The gathered data are a powerful asset to study the trends and make choices for the future. The Business Intelligence appeared in the mid-90s (the information synthesis to assist decision-making) with OLAP (On-Line Analytical Processing, a tools set for exploration, analysis and display of multidimensional data) and S-OLAP (Spatial OLAP, OLAP with spatial support). A OLAP user, unspecialized in computer sciences, does not need to know a language to handle multidimensional data, create graphics, etc. However, we consider that the OLAP data model is too rigid, because of its permanent multidimensionnal structure and because each content must have a single aggregate value. This observation is the starting point of this thesis. We propose a new paradigm of information system, able to analyze and explore multidimensional and multivalued data. To model this paradigm, we use the logical information systems (LIS) which is an information system that has common features with OLAP, especially on the data mining aspects. Our paradigm is defined by a flexible data model, an easy navigation and modular representation. We concluded this thesis by the application of this paradigm on several topics, including the exploration of geographic data.
    [bibtex-key = Allard2011PhD] [bibtex-entry]


  2. Denis Béchet, Alexandre Dikovsky, and Annie Foret. Categorial Grammars with Iterated Types form a Strict Hierarchy of k-Valued Languages. In Implementation and Application of Automata - 16th International Conference, CIAA 2011, Blois, France, volume 6807 of Lecture Notes in Computer Science, pages 42-52, 2011. Springer.
    Abstract:
    The notion of k-valued categorial grammars where a word is associated to at most k types is often used in the field of lexicalized grammars as a fruitful constraint for obtaining several properties like the existence of learning algorithms. This principle is relevant only when the classes of k-valued grammars correspond to a real hierarchy of languages. Such a property had been shown earlier for classical categorial grammars. This paper establishes the relevance of this notion when categorial grammars are enriched with iterated types.
    [bibtex-key = Foret11c] [bibtex-entry]


  3. Denis Béchet, Alexander Dikovsky, and Annie Foret. On Dispersed and Choice Iteration in Incrementally Learnable Dependency Types. In Logical Aspects of Computational Linguistics - 6th International Conference, LACL 2011, Montpellier, France, volume 6736 of Lecture Notes in Computer Science, pages 80-95, 2011. Springer. Keyword(s): Grammatical inference, Categorial grammar, Dependency grammar, Incremental learning, Iterated types..
    Abstract:
    We study learnability of Categorial Dependency Grammars (CDG), a family of categorial grammars expressing all kinds of projective, discontinuous and repeatable dependencies. For these grammars, it is known that they are not learnable from dependency structures. We propose two different ways of modelling the repeatable dependencies through iterated types and the two corresponding families of CDG which cannot distinguish between the dependencies repeatable at least K times and those repeatable any number of times. For both we show that they are incrementally learnable in the limit from dependency structures.
    [bibtex-key = Foret11b] [bibtex-entry]


  4. Denis Béchet, Alexander Dikovsky, and Annie Foret. Sur les itérations dispersées et les choix itérés pour l'apprentissage incrémental des types dans les grammaires de dépendances. In Conférence Francophone d'Apprentissage 2011 (CAP), Chambéry, France, 2011. [bibtex-key = Foret11a] [bibtex-entry]


  5. Peggy Cellier, Mireille Ducassé, and Sébastien Ferré. Exploration de traces à l'aide de fouille de données. In Atelier IC Traces numériques, connaissances et cognition, 2011. [bibtex-key = cellier2011exploration] [bibtex-entry]


  6. Peggy Cellier, Mireille Ducassé, Sébastien Ferré, and Olivier Ridoux. Multiple Fault Localization with Data Mining. In Int. Conf. on Software Engineering & Knowledge Engineering, pages 238-243, 2011. Knowledge Systems Institute Graduate School. Keyword(s): data mining, software engineering, debugging, association rules, formal concept analysis.
    Abstract:
    We have proposed an interactive fault localization method based on two data mining techniques, formal concept analysis and association rules. A lattice formalizes the partial ordering and the dependencies between the sets of program elements (e.g., lines) that are most likely to lead to program execution failures. The paper provides an algorithm to traverse that lattice starting from the most suspect places. The main contribution is that the algorithm is able to deal with any number of faults within a single execution of a test suite. In addition, a stopping criterion independent of the number of faults is provided.
    [bibtex-key = CellierDFR11seke] [bibtex-entry]


  7. Peggy Cellier, Sébastien Ferré, Mireille Ducassé, and Thierry Charnois. Partial orders and logical concept analysis to explore patterns extracted by data mining. In Int. Conf. on Conceptual Structures for Discovering Knowledge, pages 77-90, 2011. Springer. Keyword(s): data mining, partial order, selection of patterns, logical concept analysis, formal concept analysis.
    Abstract:
    Data mining techniques are used in order to discover emerging knowledge (patterns) in databases. The problem of such techniques is that there are, in general, too many resulting patterns for a user to explore them all by hand. Some methods try to reduce the number of patterns without a priori pruning. The number of patterns remains, nevertheless, high. Other approaches, based on a total ranking, propose to show to the user the top-k patterns with respect to a measure. Those methods do not take into account the user's knowledge and the dependencies that exist between patterns. In this paper, we propose a new way for the user to explore extracted patterns. The method is based on navigation in a partial order over the set of all patterns in the Logical Concept Analysis framework. It accommodates several kinds of patterns and the dependencies between patterns are taken into account thanks to partial orders. It allows the user to use his/her background knowledge to navigate through the partial order, without a priori pruning. We illustrate how our method can be applied on two different tasks (software engineering and natural language processing) and two different kinds of patterns (association rules and sequential patterns).
    [bibtex-key = cellier2011iccs] [bibtex-entry]


  8. Mireille Ducassé, Sébastien Ferré, and Peggy Cellier. Building up Shared Knowledge with Logical Information Systems. In A. Napoli and V. Vychodil, editors, Proceedings of the 8th International Conference on Concept Lattices and their Applications, pages 31-42, October 2011. INRIA. Note: ISBN 978-2-905267-78-8.
    Abstract:
    Logical Information Systems (LIS) are based on Logical Concept Analysis, an extension of Formal Concept Analysis. This paper describes an application of LIS to support group decision. A case study gathered a research team. The objective was to decide on a set of potential conferences on which to send submissions. People individually used Abilis, a LIS web server, to preselect a set of conferences. Starting from 1041 call for papers, the individual participants preselected 63 conferences. They met and collectively used Abilis to select a shared set of 42 target conferences. The team could then sketch a publication planning. The case study provides evidence that LIS cover at least three of the collaboration patterns identified by Kolfschoten, de Vreede and Briggs. Abilis helped the team to build a more complete and relevant set of information (Generate/Gathering pattern); to build a shared understanding of the relevant information (Clarify/Building Shared Understanding); and to quickly reduce the number of target conferences (Reduce/Filtering pattern).
    [bibtex-key = ducasse2011] [bibtex-entry]


  9. Sébastien Ferré and Alice Hermann. Semantic Search: Reconciling Expressive Querying and Exploratory Search. In L. Aroyo and C. Welty, editors, Int. Semantic Web Conf., LNCS 7031, pages 177-192, 2011. Springer. Keyword(s): semantic web, querying, exploratory search, expressiveness, navigation, faceted search.
    Abstract:
    Faceted search and querying are two well-known paradigms to search the Semantic Web. Querying languages, such as SPARQL, offer expressive means for searching RDF datasets, but they are difficult to use. Query assistants help users to write well-formed queries, but they do not prevent empty results. Faceted search supports exploratory search, i.e., guided navigation that returns rich feedbacks to users, and prevents them to fall in dead-ends (empty results). However, faceted search systems do not offer the same expressiveness as query languages. We introduce Query-based Faceted Search (QFS), the combination of an expressive query language and faceted search, to reconcile the two paradigms. In this paper, the LISQL query language generalizes existing semantic faceted search systems, and covers most features of SPARQL. A prototype, Sewelis (aka. Camelis 2), has been implemented, and a usability evaluation demonstrated that QFS retains the ease-of-use of faceted search, and enables users to build complex queries with little training.
    [bibtex-key = FerHer2011iswc] [bibtex-entry]


  10. S. Ferré, A. Hermann, and M. Ducassé. Combining Faceted Search and Query Languages for the Semantic Web. In C. Salinesi and O. Pastor, editors, Semantic Search over the Web (SSW) - Advanced Information Systems Engineering Workshops - CAiSE Int. Workshops, volume 83 of LNBIP 83, pages 554-563, 2011. Springer. Note: Best paper. Keyword(s): semantic web, semantic search, user interaction, faceted search, querying.
    Abstract:
    Faceted search and querying are the two main paradigms to search the Semantic Web. Querying languages, such as SPARQL, offer expressive means for searching knowledge bases, but they are difficult to use. Query assistants help users to write well-formed queries, but they do not prevent empty results. Faceted search supports exploratory search, i.e., guided navigation that returns rich feedbacks to users, and prevents them to fall in dead-ends (empty results). However, faceted search systems do not offer the same expressiveness as query languages. We introduce semantic faceted search, the combination of an expressive query language and faceted search to reconcile the two paradigms. The query language is basically SPARQL, but with a syntax that better fits in a faceted search interface. A prototype, Camelis 2, has been implemented, and a usability evaluation demonstrated that semantic faceted search retains the ease-of-use of faceted search, and enables users to build complex queries with little training.
    [bibtex-key = FerHerDuc2011ssw] [bibtex-entry]


  11. Alice Hermann, Sébastien Ferré, and Mireille Ducassé. Création et mise à jour guidées d'objets dans une base RDF(S). In Rencontres Jeunes Chercheurs en Intelligence Artificielle (RJCIA), 2011. Presses de l'Université des Antilles et de la Guyane.
    Abstract:
    La mise à jour des bases de connaissances existantes est cruciale pour tenir compte des nouvelles informations, régulièrement découvertes. Toutefois, les données actuelles du Web Sémantique sont rarement mises à jour par les utilisateurs. Les utilisateurs ne sont pas suffisament aidés lors de l'ajout et de la mise à jour d'objets. Nous proposons une approche pour aider l'utilisateur à ajouter de nouveaux objets de manière incrémentale et dynamique. Notre approche est fondée sur les Systèmes d'Information Logiques pour l'interaction utilisateur. Pour le guidage, le système cherche les objets ayant des propriétés en commun avec la description de l'objet en cours de création. Les propriétés de ces objets, non présents dans la description, servent de suggestions pour compléter la description de l'objet.
    [bibtex-key = Hermann:RJCIA:2011] [bibtex-entry]


  12. Alice Hermann, Sébastien Ferré, and Mireille Ducassé. Guided creation and update of objects in RDF(S) bases. In Mark A. Musen and Óscar Corcho, editors, Int. Conf. Knowledge Capture (K-CAP 2011), pages 189-190, 2011. ACM Press.
    Abstract:
    Updating existing knowledge bases is crucial to take into account the information that are regularly discovered. However, this is quite tedious and in practice Semantic Web data are rarely updated by users. This paper presents UTILIS, an approach to help users create and update objects in RDF(S) bases. While creating a new object, $o$, UTILIS searches for similar objects, found by applying relaxation rules to the description of $o$, taken as a query. The resulting objects and their properties serve as suggestions to expand the description of $o$.
    [bibtex-key = Hermann:KCAP:2011] [bibtex-entry]


  13. D. Legallois, Peggy Cellier, and Thierry Charnois. Calcul de réseaux phrastiques pour l analyse et la navigation textuelle. In Traitement Automatique des Langues Naturelles, 2011. Keyword(s): Sentence network, Bonds between sentences, Textual analysis, Discourse analysis.
    Abstract:
    In this paper, we present an automatic process based on text reduction introduced by Hoey. The application of that kind of approaches on large texts is difficult to do by hand. In the paper, we propose an automatic process to treat large texts. We have conducted some experiments on different kinds of texts (narrative, expositive) to show the benefits of the approach.
    [bibtex-key = legallois2011taln] [bibtex-entry]


  14. Sébastien Ferré. SQUALL: a High-Level Language for Querying and Updating the Semantic Web. Research Report, IRISA, 2011. [WWW] Keyword(s): Semantic Web, controlled natural language, query language, update language, expressiveness, Montague grammar.
    Abstract:
    Languages play a central role in the Semantic Web. An important aspect regarding their design is syntax as it plays a crucial role in the wide acceptance of the Semantic Web approach. Like for programming languages, an evolution can be observed from low-level to high-level designs. High-level languages not only allow more people to contribute by abstracting from the details, but also makes experienced people more productive, and makes the produced documents easier to share and maintain. We introduce SQUALL, a high-level language for querying and updating semantic data. It has a strong adequacy with RDF, an expressiveness very similar to SPARQL 1.1, and a controlled natural language syntax that completely abstracts from low-level notions such as bindings and relational algebra. We first give an informal presentation of SQUALL through examples, comparing it with SPARQL. We then formally define the syntax and semantics of SQUALL as a Montague grammar, and its translation to SPARQL.
    [bibtex-key = PI1985] [bibtex-entry]


  15. Sébastien Ferré, Alice Hermann, and Mireille Ducassé. Semantic Faceted Search: Safe and Expressive Navigation in RDF Graphs. Research Report, IRISA, 2011. [WWW] Keyword(s): semantic web, faceted search, query language, exploratory search, navigation, expressiveness.
    Abstract:
    Faceted search and querying are the two main paradigms to search the Semantic Web. Querying languages, such as SPARQL, offer expressive means for searching knowledge bases, but they are difficult to use. Query assistants help users to write well-formed queries, but they do not prevent empty results. Faceted search supports exploratory search, i.e., guided navigation that returns rich feedbacks to users, and prevents them to make navigation steps that lead to empty results (dead-ends). However, faceted search systems do not offer the same expressiveness as query languages. We introduce {\em semantic faceted search}, the combination of an expressive query language and faceted search to reconcile the two paradigms. The query language is basically SPARQL, but with a syntax that extends Turtle with disjunction and negation, and that better fits in a faceted search interface: LISQL. We formalize the navigation of faceted search as a navigation graph, where nodes are queries, and navigation links are query transformations. We prove that this navigation graph is {\em safe} (no dead-end), and {\em complete} (every query that is not a dead-end can be reached by navigation). That formalization itself is a contribution to faceted search. \KILL{The expressiveness of LISQL is shown comparable to SPARQL, and they are based on the same relational algebra.} A prototype, Camelis~2, has been implemented, and a usability evaluation with graduate students demonstrated that semantic faceted search retains the ease-of-use of faceted search, and enables most users to build complex queries with little training.
    [bibtex-key = PI1964] [bibtex-entry]


  16. Sébastien Ferré and Alice Hermann. Camelis2 : explorer et éditer une base RDF(S) de façon expressive et interactive. Note: Démo acceptée à la platerforme AFIA, 2011. Keyword(s): demo. [bibtex-key = demo:camelis2:ic2011] [bibtex-entry]


  17. Sébastien Ferré and Alice Hermann. Sewelis: Exploring and Editing an RDF Base in an Expressive and Interactive Way. Note: Demo accepted at the Int. Semantic Web Conf. (ISWC), 2011. Keyword(s): demo.
    Abstract:
    Query-based Faceted Search (QFS), introduced in a research paper at ISWC'11, reconciles the expressiveness of querying languages (e.g., SPARQL), and the benefits of exploratory search found in faceted search. Because of the interactive nature of QFS, which is difficult to fully render in a research paper, we feel it is important to complement it with a demonstration of our QFS prototype, Sewelis (aka. Camelis 2). An important addition to the research paper is the extension of QFS to the guided edition of RDF bases, where suggestions are based on existing data. This paper motivates our approach, shortly presents Sewelis, and announces the program of the demonstration. Screencasts of the demonstration, as well as material (program and data) to reproduce it, are available at { t http://www.irisa.fr/LIS/softwares/sewelis}.
    [bibtex-key = demo:sewelis:iswc2011] [bibtex-entry]


2010
  1. Madalina Croitoru, Sébastien Ferré, and Dickson Lukose, editors. Conceptual Structures: From Information to Intelligence, 18th International Conference on Conceptual Structures, ICCS 2010, Kuching, Sarawak, Malaysia, July 26-30, 2010. Proceedings, volume 6208 of LNCS 6208, 2010. Springer. [bibtex-key = ICCS2010] [bibtex-entry]


  2. Denis Béchet and Annie Foret. A Pregroup Toolbox for Parsing and Building Grammars of Natural Languages. Linguistic Analysis Journal, 36, 2010. Note: To appear. [bibtex-key = Foret09b] [bibtex-entry]


  3. Mireille Ducassé and Sébastien Ferré. Aide à la décision multicritère : cohérence et équité grâce à l'analyse de concepts. Revue internationale de systémique complexe et d'études relationnelles, Nouvelles Perspectives en Sciences Sociales, 5(2):181-196, Mai 2010. [bibtex-key = ducasse10] [bibtex-entry]


  4. Annie Foret. A modular and parameterized presentation of pregroup calculus. Information and Computation Journal, 208(5):395-604, May 2010. [PDF] Keyword(s): Pregroups, Lambek categorial grammars, Logic functor, Logic component, Cut elimination.
    Abstract:
    The concept of pregroup was introduced by Lambek for natural language analysis, with a close link to non-commutative linear logic. We reformulate the pregroup calculus so as to extend it by composition with other logics and calculi. The cut elimination property and the decidability property of the sequent calculus proposed in the article are shown. Properties of composed calculi are also discussed.
    [bibtex-key = Foret09d] [bibtex-entry]


  5. Pierre Allard and Sébastien Ferré. Recherche de dépendances fonctionnelles et de règles d'association avec OLAP. In S. Ben Yahia and J.-M. Petit, editors, Extraction et Gestion des Connaissances, volume RNTI-E-19 of Revue des Nouvelles Technologies de l'Information, pages 651-652, 2010. Cépaduès-Éditions. Keyword(s): association rule, functional dependency, OLAP, navigation.
    Abstract:
    Dans le domaine des bases de données, les outils de calcul de règles d'association et de dépendances fonctionnelles affichent traditionnellement les résultats sous forme d'une liste de règles, difficiles à lire. Nous proposons ici de projeter une relation d'une base de données sur un cube de données OLAP, afin d'afficher ces règles de manière plus structurée et plus intuitive. De plus, nous montrons que les liens de navigation d'OLAP peuvent aider l'utilisateur à naviguer dans ces règles produites.
    [bibtex-key = AlaFer2010] [bibtex-entry]


  6. Pierre Allard, Sébastien Ferré, and Olivier Ridoux. Discovering Functional Dependencies and Association Rules by Navigating in a Lattice of OLAP Views. In M. Kryszkiewicz and S. Obiedkov, editors, Concept Lattices and Their Applications, pages 199-210, 2010. CEUR-WS. Keyword(s): Functional Dependencies, Association Rules, FCA, OLAP, Navigation.
    Abstract:
    Discovering dependencies in data is a well-know problem in database theory. The most common rules are Functional Dependencies (FDs), Conditional Functional Dependencies (CFDs) and Association Rules (ARs). Many tools can display those rules as lists, but those lists are often too long for inspection by users. We propose a new way to display and navigate through those rules. Display is based on On-Line Analytical Processing (OLAP), presenting a set of rules as a cube, where dimensions correspond to the premises of rules. Cubes reflect the hierarchy that exists between FDs, CFDs and ARs. Navigation is based on a lattice, where nodes are OLAP views, and edges are OLAP navigation links, and guides users from cube to cube. We present an illustrative example with the help of our prototype.
    [bibtex-key = AllardFR10] [bibtex-entry]


  7. D. Bechet, A. Dikovsky, and Annie Foret. Two models of learning iterated dependencies.. In Formal Grammar, LNCS, 2010. Springer.
    Abstract:
    We study the learnability problem in the family of Categorial Dependency Grammars (CDG), a class of categorial grammars defining unlimited dependency structures. CDG satisfying a reasonable condition on the iterant dependencies are shown to be incrementally learnable in the limit.
    [bibtex-key = Foret10b] [bibtex-entry]


  8. Peggy Cellier and Thierry Charnois. Fouille de données séquentielle d'itemsets pour l'apprentissage de patrons linguistiques. In Traitement Automatique des Langues Naturelles (short paper), 2010. [bibtex-key = CellierC10] [bibtex-entry]


  9. Peggy Cellier, Thierry Charnois, and Marc Plantevit. Sequential Patterns to Discover and Characterise Biological Relations. In A. F. Gelbukh, editor, Computational Linguistics and Intelligent Text Processing (CICLing), LNCS 6008, pages 537-548, 2010. Springer. [bibtex-key = CellierCP10] [bibtex-entry]


  10. Peggy Cellier, Thierry Charnois, Marc Plantevit, and Bruno Crémilleux. Recursive Sequence Mining to Discover Named Entity Relations. In Paul R. Cohen, Niall M. Adams, and Michael R. Berthold, editors, Advances in Intelligent Data Analysis IX, 9th International Symposium (IDA), LNCS 6065, pages 30-41, 2010. Springer. [bibtex-key = CellierCPC10] [bibtex-entry]


  11. Sébastien Ferré. Conceptual Navigation in RDF Graphs with SPARQL-Like Queries. In L. Kwuida and B. Sertkaya, editors, Int. Conf. Formal Concept Analysis, LNCS 5986, pages 193-208, 2010. Springer. Keyword(s): conceptual navigation, semantic web, RDF, SPARQL, querying, navigation.
    Abstract:
    Concept lattices have been successfully used for information retrieval and browsing. They offer the advantage of combining querying and navigation in a consistent way. Conceptual navigation is more flexible than hierarchical navigation, and easier to use than plain querying. It has already been applied to formal, logical, and relational contexts, but its application to the semantic web is a challenge because of inference mechanisms and expressive query languages such as SPARQL. The contribution of this paper is to extend conceptual navigation to the browsing of RDF graphs, where concepts are accessed through SPARQL-like queries. This extended conceptual navigation is proved consistent w.r.t. the context (i.e., never leads to an empty result set), and complete w.r.t. the conjunctive fragment of the query language (i.e., every query can be reached by navigation only). Our query language has an expressivity similar to SPARQL, and has a more natural syntax close to description logics.
    [bibtex-key = Fer2010] [bibtex-entry]


  12. Annie Foret and Sébastien Ferré. On Categorial Grammars as Logical Information Systems. In L. Kwuida and B. Sertkaya, editors, Int. Conf. Formal Concept Analysis, LNCS 5986, pages 225-240, 2010. Springer. Keyword(s): logical information systems, categoarial grammar, pregroup, browsing.
    Abstract:
    We explore different perspectives on how categorial grammars can be considered as Logical Information Systems (LIS) both theoretically, and practically. Categorial grammars already have close connections with logic. We discuss the advantages of integrating both approaches. We consider more generally different ways of connecting computational linguistic data and LIS as an application of Formal Concept Analysis.
    [bibtex-key = ForFer2010] [bibtex-entry]


  13. Peggy Cellier, Mireille Ducassé, Sébastien Ferré, and Olivier Ridoux. Fouille de données pour la localisation de fautes dans les programmes, 2010. [bibtex-key = CellierDFR10] [bibtex-entry]


2009
  1. S. Ferré and S. Rudolph, editors. Formal Concept Analysis, 7th International Conference, ICFCA 2009, Darmstadt, Germany, May 21-24, 2009, Proceedings, LNCS 5548, 2009. Springer. Keyword(s): formal concept analysis. [bibtex-key = ICFCA2009] [bibtex-entry]


  2. Olivier Bedel. GEOLIS : Un Système d'information logique pour l'organisation et la recherche de données géolocalisées. PhD thesis, Thèse de l'université de Rennes 1, 22 janvier 2009. Note: Coencadrée par O. Ridoux et S. Ferré. [bibtex-key = Bedel2009PhD] [bibtex-entry]


  3. G. M. Sacco and Sébastien Ferré. Dynamic Taxonomies and Faceted Search: Theory, Practice, and Experience, volume 25 of The Information Retrieval Series, chapter 5 - Extensions to the Model, pages 113-144. Springer, 2009. Keyword(s): dynamic taxonomies, faceted search, browsing, information retrieval. [bibtex-key = Book-FIND:Chap5] [bibtex-entry]


  4. G. M. Sacco and Sébastien Ferré. Dynamic Taxonomies and Faceted Search: Theory, Practice, and Experience, volume 25 of The Information Retrieval Series, chapter 9 - Applications and Experiences, pages 263-302. Springer, 2009. Keyword(s): dynamic taxonomies, faceted search, browsing, information retrieval. [bibtex-key = Book-FIND:Chap9] [bibtex-entry]


  5. G. M. Sacco, Sébastien Ferré, and Y. Tzitzikas. Dynamic Taxonomies and Faceted Search: Theory, Practice, and Experience, volume 25 of The Information Retrieval Series, chapter 3 - Comparison with Other Techniques, pages 35-74. Springer, 2009. Keyword(s): dynamic taxonomies, faceted search, browsing, information retrieval. [bibtex-key = Book-FIND:Chap3] [bibtex-entry]


  6. G. M. Sacco, Y. Tzitzikas, and Sébastien Ferré. Dynamic Taxonomies and Faceted Search: Theory, Practice, and Experience, volume 25 of The Information Retrieval Series, chapter 8 - System Implementation, pages 215-262. Springer, 2009. Keyword(s): dynamic taxonomies, faceted search, browsing, information retrieval. [bibtex-key = Book-FIND:Chap8] [bibtex-entry]


  7. M. Stefaner, Sébastien Ferré, S. Perugini, J. Koren, and Y. Zhang. Dynamic Taxonomies and Faceted Search: Theory, Practice, and Experience, volume 25 of The Information Retrieval Series, chapter 4 - User Interface Design, pages 75-112. Springer, 2009. Keyword(s): dynamic taxonomies, faceted search, browsing, information retrieval. [bibtex-key = Book-FIND:Chap4] [bibtex-entry]


  8. Sébastien Ferré. Camelis: a logical information system to organize and browse a collection of documents. Int. J. General Systems, 38(4):379-403, 2009. Keyword(s): information retrieval, browsing, logical concept analysis, annotation.
    Abstract:
    Since the arrival of digital cameras, many people are faced to the challenge of organizing and browsing the overwhelming flood of photos their life produces. The same is true for all sorts of documents, e.g.~emails, audio files. Existing systems either let users fill query boxes without any assistance, or drive them through rigid navigation structures (e.g., hierarchies); or they do not let users put annotations on their documents, even when this would support the organization and retrieval of any documents on customized criteria. We present a tool, {\sc Camelis}, that offers users with an organization that is dynamically computed from documents and their annotations. {\sc Camelis} is designed along the lines of Logical Information Systems (LIS), which are founded on logical concept analysis. Hence, (1) an expressive language can be used to describe photos and query the collection, (2) manual and automatic annotations can be smoothly integrated, and (3) expressive querying and flexible navigation can be mixed in a same search and in any order. This presentation is illustrated on a real collection of more than 5,000 photos.
    [bibtex-key = Fer2009] [bibtex-entry]


  9. B. Morin, Ludovic Mé, Hervé Debar, and Mireille Ducassé. A Logic-based model to support alert correlation in intrusion detection. Information Fusion, 10(4):285-299, October 2009. Keyword(s): Intrusion detection, Alert correlation, Data model.
    Abstract:
    Managing and supervising security in large networks has become a challenging task, as new threats and flaws are being discovered on a daily basis. This requires an in depth and up-to-date knowledge of the context in which security-related events occur. Several tools have been proposed to support security operators in this task, each of which focuses on some specific aspects of the monitoring. Many alarm fusion and correlation approaches have also been investigated. However, most of these approaches suffer from two major drawbacks. First, they only take advantage of the information found in alerts, which is not sufficient to achieve the goals of alert correlation, that is to say to reduce the overall amount of alerts, while enhancing their semantics. Second, these techniques have been designed on an ad hoc basis and lack a shared data model that would allow them to reason about events in a cooperative way. In this paper, we propose a federative data model for security systems to query and assert knowledge about security incidents and the context in which they occur. This model constitutes a consistent and formal ground to represent information that is required to reason about complementary evidences, in order to confirm or invalidate alerts raised by intrusion detection systems.
    [bibtex-key = morin09] [bibtex-entry]


  10. Denis Béchet and Annie Foret. (PPQ) : a pregroup parser using majority composition. In Proc. Parsing with Categorial Grammars, ESSLLI workshop, in Bordeaux, France, 2009. [PDF] Keyword(s): parser, pregroups, Lambek categorial grammars, parsing software, XML data.
    Abstract:
    Pregroup grammars are a mathematical formalism in the spirit of categorial grammars. They are close to logical formalism like Lambek calculus but have a polynomial parsing algorithm. The paper presents a parser based on pregroup gram- mar that uses a tabular approach based on majority partial composition.
    [bibtex-key = Foret09a] [bibtex-entry]


  11. Denis Béchet and Annie Foret. Une boîte à outils pour développer et utiliser les grammaires de prégroupe. In journée ATALA : Quels analyseurs syntaxiques pour le français ?, 2009. [PDF]
    Abstract:
    Pregroup grammars are a formalism in the spirit of categorial grammars and the Lambek calculus. In contrast to the latter, their parsing is polynomial. We present in this article a toolbox that contains a parser and a set of programs for the construction and use of grammars, in particular for French.
    [bibtex-key = Foret09c] [bibtex-entry]


  12. Peggy Cellier, Mireille Ducassé, Sébastien Ferré, and Olivier Ridoux. DeLLIS: A Data Mining Process for Fault Localization. In Int. Conf. Software Engineering (SEKE), pages 432-437, 2009. Knowledge Systems Institute Graduate School. Keyword(s): data mining, fault localization, software. [bibtex-key = CDFR2009] [bibtex-entry]


  13. Mireille Ducassé. (C)LP Tracing and Debugging. In Patricia M. Hill and David Scott Warren, editors, 25th International Conference on Logic Programming, ICLP 2009, volume 5649 of Lecture Notes in Computer Science, pages 38, July 2009. Springer. [bibtex-key = ducasse09c] [bibtex-entry]


  14. Mireille Ducassé and Sébastien Ferré. Aide à la décision multicritère : cohérence et équité grâce à l'analyse de concepts. In Modèles et Apprentissage en Sciences Humaines et Sociales, Juin 2009. Keyword(s): Décision multicritère, aide à la décision, choix social, analyse formelle de concepts, systèmes d'information logiques, étude de cas.
    Abstract:
    De nombreuses décisions sont prises en commission, par exemple pour affecter des ressources. Les critères de décision sont difficiles à exprimer et la situation globale est en général trop complexe pour que les participants puissent l'appréhender pleinement. Dans cet article, nous décrivons un processus de décision où l'analyse de concepts est utilisée pour faire face à ces problèmes. Grâce à l'analyse de concepts, les personnes fair play ont la possibilité d'être équitables envers les candidats et de faire preuve de cohérence dans leurs jugements sur toute la durée de la réunion.
    [bibtex-key = ducasse09b] [bibtex-entry]


  15. Sébastien Ferré. Efficient Browsing and Update of Complex Data Based on the Decomposition of Contexts. In S. Rudolph, F. Dau, and S. O. Kuznetsov, editors, Int. Conf. Conceptual Structures, LNCS 5662, pages 159-172, 2009. Springer. Keyword(s): software component, logical concept analysis, browsing, update.
    Abstract:
    Formal concept analysis is recognized as a good paradigm for browsing data sets. Besides browsing, update and complex data are other important aspects of information systems. To have an efficient implementation of concept-based information systems is difficult because of the diversity of complex data and the computation of conceptual structures, but essential for the scalability to real-world applications. We propose to decompose contexts into simpler and specialized components: logical context functors. We demonstrate this allows for scalable implementations, updatable ontologies, and richer navigation structures, while retaining genericity.
    [bibtex-key = Fer2009b] [bibtex-entry]


  16. Sébastien Ferré. Navigating the Semantic Web with Logical Information Systems. internal publication 1934, Institut de Recherche en Informatique et Systèmes Aléatoires (IRISA), August 2009. [WWW] Keyword(s): logical information systems, semantic web, navigation, querying.
    Abstract:
    Exploratory search is about browsing and understanding an information base. It requires a more interactive process than retrieval search, where the user sends a query and the system returns answers. Logical Information Systems (LIS) support exploratory search by guiding users in the construction of queries, and by giving summaries of query results. The contribution of this paper is to adapt and extend LIS to the Semantic Web. We define a summarization and navigation data structure that provides rich views over data, and guides users from view to view. This guiding is proved consistent (i.e., never leads to an empty result set), and complete (i.e., every query can be reached by navigation only). The query language covers a large fragment of SPARQL, and is more concise by using a syntax close to description logics. Our approach could be implemented on top of existing tools for storing, reasoning and querying.
    [bibtex-key = Fer2009c] [bibtex-entry]


  17. Mireille Ducassé. Cours de conception participative dans la formation d'ingénieur en informatique. Poster aux Rencontres ``Pédagogie et Formations d'Ingénieurs'', Mars 2009. Note: INSA de Toulouse.
    Abstract:
    On entend souvent des plaintes au sujet d'applications informatiques inadaptées à leurs utilisateurs. Nous pensons que les informaticiens doivent être sensibilisés à la prise en compte des utilisateurs dans la conception et le développement de logiciels. Cela ne relève pas de techniques informatiques, mais c'est néanmoins suffisement délicat et complexe pour justifier un cours. Au département informatique de l'INSA de Rennes, les étudiants réalisent une analyse par conception participative d'un produit grand public. C'est un travail par groupe et collaboratif, tous les groupes travaillent sur le même produit mais avec des utilisateurs différents. Le cours est également sa propre illustration. L'avis des étudiants est systématiquement collecté et le cours est en constante évolution.
    [bibtex-key = ducasse09] [bibtex-entry]


2008
  1. Olivier Ridoux and Gilles Lesventes. Calculateurs, calculs, calculabilité, Sciences Sup. Dunod, 2008. [bibtex-key = ccc] [bibtex-entry]


  2. Peggy Cellier. DeLLIS : Débogage de programmes par Localisation de fautes avec un Système d'Information Logique. PhD thesis, Thèse de l'université de Rennes 1, 05 décembre 2008. Note: Coencadrée par M. Ducassé, S. Ferré et O. Ridoux. [bibtex-key = Cellier2008PhD] [bibtex-entry]


  3. Tristan Denmat. Contraintes et abstractions pour la génération automatique de données de test. PhD thesis, Thèse de l'INSA de Rennes, 05 juin 2008. Note: Coencadrée par M. Ducassé et Anaud Gotlieb. [bibtex-key = Denmat2008PhD] [bibtex-entry]


  4. D. Bechet, A. Dikovsky, Annie Foret, and Emmanuelle Garel. Introduction of option and iteration into pregroup grammars. In Computational Algebraic Approaches to Natural Language. Polimetrica Publisher, Italy, pp. 85-108.. 2008. Keyword(s): Pregroups, Lambek Categorial Grammars, Categorial Dependency Grammar.
    Abstract:
    We discuss the relationship between pregroups and dependency grammars. Conventional pregroup grammars do not formally account for optionality and for iteration. We introduce Gentzen-style rules to take care of two new operations, and an equivalent rewriting system. The extended pregroup calculus enjoys several properties shared with traditional dependency grammars, yet does not significantly expand the polynomial complexity of the syntactic analysis on the pregroup grammar.
    [bibtex-key = Foret08b] [bibtex-entry]


  5. Olivier Bedel, Sébastien Ferré, Olivier Ridoux, and Erwan Quesseveur. GEOLIS: A Logical Information System for Geographical Data. Revue Internationale de Géomatique, 17(3-4):371-390, 2008. [PDF] Keyword(s): logical information system, geographical data, navigation.
    Abstract:
    Today, the thematic layer is still the prevailling structure in geomatics for handling geographical information. However, the layer model is rigid: it implies partitionning geographical data in predefined categories and using the same description schema for all elements of a layer. Recently, Logical Information Systems (LIS) introduced a new paradigm for information management and retrieval. Using LIS, we propose a more flexible organisation of vectorial geographical data at a thiner level since it is centered on the geographical feature. LIS do not rely on a hierarchical organisation of information, and enable to tightly combine querying and navigation. In this article, we present the use of LIS to handle geographical data. In particular, we detail a data model for geographical features and the corresponding querying and navigation model. These models have been implemented in the GEOLIS prototype, which has been used to lead experiments on real data.
    [bibtex-key = BFRQ2008] [bibtex-entry]


  6. Peggy Cellier, Sébastien Ferré, Olivier Ridoux, and Mireille Ducassé. A Parameterized Algorithm to Explore Formal Contexts with a Taxonomy. Int. J. Foundations of Computer Science (IJFCS), 19(2):319-343, 2008. Keyword(s): algorithm, concept lattice, taxonomy.
    Abstract:
    Formal Concept Analysis (FCA) is a natural framework to learn from examples. Indeed, learning from examples results in sets of frequent concepts whose extent contains mostly these examples. In terms of association rules, the above learning strategy can be seen as searching the premises of rules where the consequence is set. In its most classical setting, FCA considers attributes as a non-ordered set. When attributes of the context are partially ordered to form a taxonomy, Conceptual Scaling allows the taxonomy to be taken into account by producing a context completed with all attributes deduced from the taxonomy. The drawback, however, is that concept intents contain redundant information. In this article, we propose a parameterized algorithm, to learn rules in the presence of a taxonomy. It works on a non-completed context. The taxonomy is taken into account during the computation so as to remove all redundancies from intents. Simply changing one of its operations, this parameterized algorithm can compute various kinds of concept-based rules. We present instantiations of the parameterized algorithm to learn rules as well as to compute the set of frequent concepts.
    [bibtex-key = CFRD2008] [bibtex-entry]


  7. Ludovic Langevine and Mireille Ducassé. Design and Implementation of a Tracer Driver: Easy and Efficient Dynamic Analyses of Constraint Logic Programs. Theory and Practice of Logic Programming, Cambridge University Press, 8(5-6), Sep-Nov 2008. [WWW] [bibtex-key = langevine08] [bibtex-entry]


  8. Pierre Allard and Sébastien Ferré. Dynamic Taxonomies for the Semantic Web. In A. M. Tjoa and R. R. Wagner, editors, DEXA Int. Work. Dynamic Taxonomies and Faceted Search (FIND), pages 382-386, 2008. IEEE Computer Society. [PDF] Keyword(s): semantic web, ontologies, logical information system, dynamic taxonomies.
    Abstract:
    The semantic web aims at enabling the web to understand and answer the requests from people and machines. It relies on several standards for representing and reasoning about web contents. Among them, the Web Ontology Language (OWL) is used to define ontologies, i.e. knowledge bases, and is formalized with description logics. In this paper, we demonstrate how dynamic taxonomies and their benefits can be transposed to browse OWL~DL ontologies. We only assume the ontology has an assertional part, i.e. defines objects and not only concepts. The existence of relations between objects in OWL leads us to define new navigation modes for crossing these relations. A prototype, Odalisque, has been developed on top of well-known tools for the semantic web.
    [bibtex-key = AlaFer2008] [bibtex-entry]


  9. D. Bechet, A. Dikovsky, Annie Foret, and Emmanuelle Garel. Optional and Iterated Types for Pregroup Grammars. In Int. Conf. Language and Automata Theory and Applications (LATA), 2008. Note: Accepted for publication. Keyword(s): Pregroups, Lambek Categorial Grammars, Categorial Dependency Grammar.
    Abstract:
    Pregroup grammars are a context-free grammar formalism which may be used to describe the syntax of natural languages. However, this formalism is not able to easily define types corresponding to optional or iterated arguments like an optional complement of a verb or a sequence of its adverbial modifiers. This paper introduces two constructions that %% solve this issue. make up for this deficiency.
    [bibtex-key = Foret08a] [bibtex-entry]


  10. Olivier Bedel, Sébastien Ferré, and Olivier Ridoux. Handling Spatial Relations in Logical Concept Analysis To Explore Geographical Data. In R. Medina and S. Obiedkov, editors, Int. Conf. Formal Concept Analysis, LNAI 4933, pages 241-257, 2008. Springer. [PDF] Keyword(s): spatial relations, concept analysis, logic, geographical data, data retrieval.
    Abstract:
    Because of the expansion of geo-positioning tools and the democratization of geographical information, the amount of geo-localized data that is available around the world keeps increasing. So, the ability to efficiently retrieve informations in function of their geographical facet is an important issue. In addition to individual properties such as position and shape, spatial relations between objects are an important criteria for selecting and reaching objects of interest: e.g., given a set of touristic points, selecting those having a nearby hotel or reaching the nearby hotels. In this paper, we propose Logical Concept Analysis (LCA) and its handling of relations for representing and reasoning on various kinds of spatial relations: e.g., Euclidean distance, topological relations. Furthermore, we present an original way of navigating in geolocalized data, and compare the benefits of our approach with traditional Geographical Information Systems (GIS).
    [bibtex-key = BedFerRid2008] [bibtex-entry]


  11. Peggy Cellier. Formal concept analysis applied to fault localization. In Robby, editor, International Conference on Software Engineering (ICSE 2008) Companion, pages 991-994, 2008. ACM. [PDF] [bibtex-key = icse08] [bibtex-entry]


  12. Peggy Cellier, Mireille Ducassé, Sébastien Ferré, and Olivier Ridoux. Formal Concept analysis enhances Fault Localization in Software. In R. Medina and S. Obiedkov, editors, Int. Conf. Formal Concept Analysis, LNAI 4933, pages 273-288, 2008. Springer. [PDF] Keyword(s): fault localization, formal concept analysis.
    Abstract:
    Recent work in fault localization crosschecks traces of correct and failing execution traces. The implicit underlying technique is to search for association rules which indicate that executing a particular source line will cause the whole execution to fail. This technique, however, has limitations. In this article, we first propose to consider more expressive association rules where several lines imply failure. We then propose to use Formal Concept Analysis (FCA) to analyze the resulting numerous rules in order to improve the readability of the information contained in the rules. The main contribution of this article is to show that applying two data mining techniques, association rules and FCA, produces better results than existing fault localization techniques.
    [bibtex-key = CDFR2008a] [bibtex-entry]


  13. Mireille Ducassé and Sébastien Ferré. Fair(er) and (almost) serene committee meetings with Logical and Formal Concept Analysis. In P. Eklund and O. Haemmerlé, editors, Proceedings of the International Conference on Conceptual Structures, LNAI 5113, pages 217-230, July 2008. Springer. [PDF]
    Abstract:
    In academia, many decisions are taken in committee, for example to hire people or to allocate resources. Genuine people often leave such meetings quite frustrated. Indeed, it is intrinsically hard to make multi-criteria decisions, selection criteria are hard to express and the global picture is too large for participants to embrace it fully. In this article, we describe a recruiting process where logical concept analysis and formal concept analysis are used to address the above problems. We do not pretend to totally eliminate the arbitrary side of the decision. We claim, however, that, thanks to concept analysis, genuine people have the possibility to 1) be fair with the candidates, 2) make a decision adapted to the circumstances, 3) smoothly express the rationales of decisions, 4) be consistent in their judgements during the whole meeting, 5) vote (or be arbitrary) only when all possibilities for consensus have been exhausted, and 6) make sure that the result, in general a total order, is consistent with the partial orders resulting from the multiple criteria.
    Annotation:
    taux acceptation 19/70 = 27\%
    [bibtex-key = ducasse08] [bibtex-entry]


  14. Sébastien Ferré. Agile Browsing of a Document Collection with Dynamic Taxonomies. In A. M. Tjoa and R. R. Wagner, editors, DEXA Int. Work. Dynamic Taxonomies and Faceted Search (FIND), pages 377-381, 2008. IEEE Computer Society. [PDF] Keyword(s): browsing, navigation, logical information system, dynamic taxonomies.
    Abstract:
    Dynamic taxonomies and faceted search are increasingly used to organize and browse document collections. The main function of dynamic taxonomies is to start with the full collection, and zoom-in to a small enough subset of items for direct inspection. In this paper, we present other navigation modes than zoom-in for less directed and more exploratory browsing of a document collection. The presented navigation modes are zoom-out, shift, pivot, and querying by examples. These modes correspond to query transformations, and make use of boolean operators. Therefore, the current focus is always clearly specified by a query.
    [bibtex-key = Fer2008] [bibtex-entry]


2007
  1. D. Bechet, A. Foret, and I. Tellier. Learnability of Pregroup Grammars. Studia Logica, 87(2-3), 2007. Keyword(s): Learning from positive examples, Pregroup grammars, Computational linguistics, parsing, Categorial Grammars, constraints.
    Abstract:
    This paper investigates the learnability by positive examples in the sense of Gold of Pregroup Grammars. In a first part, Pregroup Grammars are presented and a new parsing strategy is prop osed. Then, theoretical learnability and non-learnability results for subclasses of Pregroup Grammars are proved. In the last two parts, we focus on learning Pregroup Grammars from a sp ecial kind of input called feature-tagged examples. A learning algorithm based on the parsing strategy presented in the first part is given. Its validity is proved and its prop erties are examplified.
    [bibtex-key = Foret07a] [bibtex-entry]


  2. D. Béchet, R. Bonato, A. Dikovsky, A. Foret, Y. Le Nir, E. Moreau, C. Retoré, and I. Tellier. Modèles algorithmiques de l'acquisition de la syntaxe : concepts et méthodes, résultats et problèmes. Recherches linguistiques de Vincennes, 2007. Note: Vol. 37, Presses Universitaires de Vincennes. [PDF] Keyword(s): Acquisition syntaxique, inférence grammaticale, grammaires catégorielles, modèle de Gold, ressources syntaxiques (syntax learning, grammatical inference, categorial grammars, Gold's model, syntactical resources).
    Abstract:
    Dans cet article, nous présentons nos résultats récents concernant l'apprentissage de la syntaxe des langues naturelles, en adoptant le point de vue de l'inférence grammaticale symbolique. L'objectif est d'identifier à partir d'exemples, dans une classe de grammaires connue à l'avance, une grammaire particulière qui engendre les dits exemples. Le modèle de Gold fixe les conditions et le critère de réussite d'une telle entreprise : quand un algorithme produisant une grammaire candidate existe-t-il ? quelle structure doivent contenir les exemples : suites de mots, suites de mots étiquetés, arbres d'analyse ? D'un point de vue théorique, nos résultats établissent l'apprenabilité ou la non-apprenabilité de certaines classes de grammaires catégorielles. En pratique, nos résultats permettent aussi d'acquérir automatiquement des ressources syntaxiques à partir de données réelles. Au final, nous discutons de l'intérêt de cette approche pour modéliser l'acquisition de leur langue naturelle par les enfants ainsi que pour construire automatiquement des grammaires électroniques à partir de corpus. In this paper, we present our recent results on the acquistion of the syntax of natural languages, from the point of view of the theory of grammatical inference. Given a class of possible grammars, the objective is to identify, from a set of positive examples, a grammar in the class which produces the examples. The Gold model formalises the learning process and gives stringent criteria of its success: when does there exist an algorithm producing a target grammar ? what kind of structure should the examples have (strings of words, strings of tagged words, trees) ? From a theoretical point of view, our results establish the learnability or the unlearnability of various classes of categorial grammars. From a practical perspective, these results enable the extraction of syntactic information from real data. Finally, we discuss the interest of this approach for modelling child language acquisition and for automated induction of grammars from corpora.
    [bibtex-key = Foret07b] [bibtex-entry]


  3. Olivier Bedel, Sébastien Ferré, and Olivier Ridoux. Exploring a Geographical Dataset with GEOLIS. In DEXA Work. Advances in Conceptual Knowledge Engineering (ACKE), pages 540-544, 2007. IEEE Computer Society. [PDF] Keyword(s): logical information system, geographical information system, information retrieval, spatial logic.
    Abstract:
    Geographical data are mainly structured in layers of information. However, this model of organisation is not convenient for navigation inside a dataset, and so limits geographical data exploration to querying. We think information retrieval could be made easier in GIS by the introduction of a navigation based on geographical object properties. For this purpose, we propose a prototype, GEOLIS1, which tightly combines querying and navigation in the search process of geographical data. GEOLIS relies on Logical Information Systems (LIS), which are based on Formal Concept Analysis (FCA) and logics. In this paper, we detail data organisation and navigation process in GEOLIS. We also present the results of an experimentation led on a real dataset.
    [bibtex-key = BedFerRid2007a] [bibtex-entry]


  4. D. Béchet and A. Foret. Fully Lexicalized Pregroup Grammars. In Proceedings of WOLLIC 2007, volume LNCS 4576, pages 12-25, 2007. Springer. [PDF] Keyword(s): Pregroups, Lambek Categorial Grammars, Simulation.
    Abstract:
    Pregroup grammars are a context-free grammar formalism introduced as a simplification of Lambek calculus. This formalism is interesting for several reasons: the syntactical properties of words are specified by a set of types like the other type-based grammar formalisms ; as a logical model, compositionality is easy ; a polytime parsing algorithm exists. However, this formalism is not completely lexicalized because each pregroup grammar is based on the free pregroup built from a set of primitive types together with a partial order, and this order is not lexical information. In fact, only the pregroup grammars that are based on primitive types with an order that is equality can be seen as fully lexicalized. We show here how we can transform, using a morphism on types, a particular pregroup grammar into another pregroup grammar that uses the equality as the order on primitive types. This transformation is at most quadratic in size (linear for a fixed set of primitive types), it preserves the parse structures of sentences and the number of types assigned to a word.
    [bibtex-key = Foret07d] [bibtex-entry]


  5. Peggy Cellier, Sébastien Ferré, Olivier Ridoux, and Mireille Ducassé. A Parameterized Algorithm for Exploring Concept Lattices. In S.O. Kuznetsov and S. Schmidt, editors, Int. Conf. Formal Concept Analysis, LNAI 4390, pages 114-129, 2007. Springer. Keyword(s): concept analysis, algorithm, taxonomy.
    Abstract:
    Formal Concept Analysis (FCA) is a natural framework for learning from positive and negative examples. Indeed, learning from examples results in sets of frequent concepts whose extent contains only these examples. In terms of association rules, the above learning strategy can be seen as searching the premises of exact rules where the consequence is fixed. In its most classical setting, FCA considers attributes as a non-ordered set. When attributes of the context are ordered, Conceptual Scaling allows the related taxonomy to be taken into account by producing a context completed with all attributes deduced from the taxonomy. The drawback, however, is that concept intents contain redundant information. In this article, we propose a parameterized generalization of a previously proposed algorithm, in order to learn rules in the presence of a taxonomy. The taxonomy is taken into account during the computation so as to remove all redundancies from intents. Simply changing one component, this parameterized algorithm can compute various kinds of concept-based rules. We present instantiations of the parameterized algorithm for learning positive and negative rules.
    [bibtex-key = CFRD2007] [bibtex-entry]


  6. Tristan Denmat, Arnaud Gotlieb, and Mireille Ducassé. An Abstract Interpretation-based Combinator for Modelling While Loops in Constraint Programming. In C. Bessière, editor, Int. Conf. Principles and Practice of Constraint Programming (CP), LNCS 4741, September 2007. Springer-Verlag. [bibtex-key = denmat07] [bibtex-entry]


  7. Tristan Denmat, Arnaud Gotlieb, and Mireille Ducassé. Improving Constraint-Based Testing with Dynamic Linear Relaxations. In K. Goseva-Popstojanova and P. Runeson, editors, Int. Symp. Software Reliability Engineering (ISSRE), November 2007. IEEE Press. [bibtex-key = denmat07b] [bibtex-entry]


  8. Pierre Deransart, Mireille Ducassé, and Gérard Ferrand. Une sémantique observationnelle du modèle des boîtes pour la résolution de programmes logiques. In F. Fages, editor, Actes de Troisièmes Journées Francophones de Programmation par Contraintes, 2007. HAL : http://hal.inria.fr/JFPC07. [bibtex-key = deransart07] [bibtex-entry]


  9. Sébastien Ferré. CAMELIS: Organizing and Browsing a Personal Photo Collection with a Logical Information System. In J. Diatta, P. Eklund, and M. Liquière, editors, Int. Conf. Concept Lattices and Their Applications, volume 331 of CEUR Workshop Proceedings ISSN 1613-0073, pages 112-123, 2007. [PDF] Keyword(s): logical information system, photo collection, organization, information retrieval.
    Abstract:
    Since the arrival of digital cameras, many people are faced to the challenge of organizing and retrieving the overwhelming flow of photos their life produces. Most people put no metadata on their photos, and we believe this is because existing tools make a very limited use of them. We present a tool, Camelis, that offers users with an organization of photos that is dynamically computed from the metadata, making worthwhile the effort to produce it. Camelis is designed along the lines of Logical Information Systems (LIS), which are founded on logical concept analysis. Hence, (1) an expressive language can be used to describe photos and query the collection, (2) manual and automatic metadata can be smoothly integrated, and (3) expressive querying and flexible navigation can be mixed in a same search and in any order. This presentation is illustrated by experiences on a real collection of more than 5000 photos.
    [bibtex-key = Fer2007b] [bibtex-entry]


  10. Sébastien Ferré. The Efficient Computation of Complete and Concise Substring Scales with Suffix Trees. In S.O. Kuznetsov and S. Schmidt, editors, Int. Conf. Formal Concept Analysis, LNAI 4390, pages 98-113, 2007. Springer. Keyword(s): concept analysis, logic, string, suffix tree.
    Abstract:
    Strings are an important part of most real application multi-valued contexts. Their conceptual treatment requires the definition of {\em substring scales}, i.e., sets of relevant substrings, so as to form informative concepts. However these scales are either defined by hand, or derived in a context-unaware manner (e.g., all words occuring in string values). We present an efficient algorithm based on suffix trees that produces complete and concise substring scales. Completeness ensures that every possible concept is formed, like when considering the scale of all substrings. Conciseness ensures the number of scale attributes (substrings) is less than the cumulated size of all string values. This algorithm is integrated in Camelis, and illustrated on the set of all ICCS paper titles.
    [bibtex-key = Fer2007a] [bibtex-entry]


  11. Sébastien Ferré and Olivier Ridoux. Logical Information Systems: from Taxonomies to Logics. In DEXA Work. Dynamic Taxonomies and Faceted Search (FIND), pages 212-216, 2007. IEEE Computer Society. [PDF] Keyword(s): logical information system, taxonomy, logic.
    Abstract:
    Dynamic taxonomies have been proposed as a solution for combining querying and navigation, offering both expressivity and interactivity. Navigation is based on the filtering of a multidimensional taxonomy w.r.t. query answers, which helps users to focus their search. We show that properties that are commonly used only in queries can be integrated in taxonomies, and hence in navigation, by the use of so-called logics. Hand-designed taxonomies and concrete domains (e.g., dates, strings) can be combined so as to form complex taxonomies. For instance, valued attributes can be handled, and different roles between documents and locations can be distinguished. Logical Information Systems (LIS) are characterized by the combination of querying and navigation, and the systematic use of logics.
    [bibtex-key = FerRid2007] [bibtex-entry]


  12. A. Foret. Pregroup Calculus as a Logical Functor. In Proceedings of WOLLIC 2007, volume LNCS 4576, 2007. Springer. [PDF] Keyword(s): Pregroups, Lambek Categorial Grammars, Logic Functor, Cut Elimination.
    Abstract:
    The concept of pregroup was introduced by Lambek for natural language analysis, with a close link to non-commutative linear logic. We reformulate the pregroup calculus so as to extend it by composition with other logics and calculii.The cut elimination property and the decidabilityproperty of the sequent calculus proposed in the article are shown.Properties of composed calculii are also discussed.
    [bibtex-key = Foret07c] [bibtex-entry]


2006
  1. Elvis Tombini. Amélioration du diagnostic en détection d'intrusions : étude et application d'une combinaison de méthodes de détection d'intrusions comportementale et par scénarios. PhD thesis, INSA de Rennes, septembre 2006. Note: M. Ducassé, H. Debar et L. Mé, directeurs de thèse. [bibtex-key = Tombini06] [bibtex-entry]


  2. Denis Bechet and Annie Foret. k-Valued Non-Associative Lambek Grammars are learnable from Generalized Functor-Argument Structures. Journal of Theoretical Computer Science, 2006. Note: Vol. 355, 2 (extended version). [PDF] Keyword(s): Grammatical inference, Categorial grammars, Non-associative Lambek calculus, Learning from positive examples, Model of Gold.
    Abstract:
    This paper is concerned with learning categorial grammars from positive examples in the model of Gold. Functor-argument structures (written FA) are usual syntactical decompositions of sentences in sub-components distinguishing the functional parts from the argument parts defined in the case of classical categorial grammars also known as AB-grammars. In the case of non-associative type-logical grammars, we propose a similar notion that we call generalized functor-argument structures and we show that these structures capture the essence of non-associative Lambek (NL) calculus without product. We show that (i) rigid and k-valued non-associative Lambek (NL without product) grammars are learnable from generalized functor-argument structured sentences. We also define subclasses of k-valued grammars in terms of arity. We first show that (ii) for each k and each bound on arity the class of FA-arity bounded k-valued NL languages of FA structures is finite and (iii) that FA-arity bounded k-valued NL grammars are learnable both from strings and from FA structures as a corollary. Result (i) is obtained from (ii); this learnability result (i) is interesting and surprising when compared to other results: in fact we also show that (iv) this class has infinite elasticity. Moreover, these classes are very close to classes like rigid associative Lambek grammars learned from natural deduction structured sentences (that are different and much richer than FA or generalized FA) or to k-valued non-associative Lambek grammars unlearnable from strings or even from bracketed strings. Thus, the class of k-valued non-associative Lambek grammars learned from generalized functor-argument sentences is at the frontier between learnable and unlearnable classes of languages.
    [bibtex-key = Foret06tcs] [bibtex-entry]


  3. Sébastien Ferré and R. D. King. Finding Motifs in Protein Secondary Structure for Use in Function Prediction. Journal of Computational Biology, 13(3):719-731, 2006. Keyword(s): functional genomics, protein secondary structure, flexible motifs, dichotomic search algorithm. [bibtex-key = FerKin2006] [bibtex-entry]


  4. Benjamin Sigonneau and Olivier Ridoux. Indexation multiple et automatisée de composants logiciels. Technique et Science Informatiques, 2006. Note: À paraître. [PDF]
    Abstract:
    Software reuse requires that programmers be able to locate reusable components in software repositories. We propose that a general information retrieval framework, which is able to combine arbitrary indexing schemes and called Logical Information Systems, is applied to querying in software repositories. As an illustration, indexing of methods in a package is studied, and three indexing schemes are presented in this framework: a formal scheme, a semi-formal one, and an informal one. The formal one captures object-orientation by combining type isomorphism axioms and inheritance relations. The semi-formal scheme captures naming conventions and the informal one captures keywords in comments. Theory of the formal methods and details on the experiments are presented.
    Annotation:
    Numéro thématique AFADL'04
    [bibtex-key = tsi-indexation-composants] [bibtex-entry]


  5. Olivier Bedel, Sébastien Ferré, Olivier Ridoux, and Erwan Quesseveur. GEOLIS: A Logical Information System for Geographical Data. In Int. Conf. Spatial Analysis and GEOmatics - SAGEO 2006, 2006. [PDF] Keyword(s): logical information systems, geographical data, navigation, querying.
    Abstract:
    Today, the thematic layer is still the prevailling structure in geomatics for handling geographical information. However, the layer model is rigid: it implies partitionning geographical data in predefined categories and using the same description schema for all elements of a layer. Recently, Logical Information Systems (LIS) introduced a new paradigm for information management and retrieval. Using LIS, we propose a more flexible organisation of vectorial geographical data at a thiner level since it is centered on the geographical feature. LIS does not rely on a hierarchical organisation of information, and enable to tightly combine querying and navigation in a same search. In this article, we present a work in progress about the use of LIS model to handle geographical data. In particular, we detail a data model for geographical features and the corresponding querying and navigation model. These models have been implemented in the GEOLIS prototype, which has been used to lead experiments with real data.
    [bibtex-key = BFRQ2006] [bibtex-entry]


  6. Olivier Bedel, Olivier Ridoux, and Erwan Quesseveur. Combining Logical Information System and OpenGIS Tools for Geographical Data Exploration. In Int. Conf. Free and OpenSource Sofware for Geoinformatics, september 2006. [WWW] Keyword(s): logical information system, GIS, graphical data. [bibtex-key = BedRidQue2006] [bibtex-entry]


  7. Peggy Cellier, Sébastien Ferré, Olivier Ridoux, and Mireille Ducassé. An Algorithm to Find Frequent Concepts of a Formal Context with Taxonomy. In S. Ben Yahia and E. Mephu Nguifo, editors, Int. Conf. Concept Lattices and Their Applications, LNAI 4932, pages 243-248, 2006. Springer. [PDF] Keyword(s): concept analysis, algorithm, taxonomy.
    Abstract:
    Formal Concept Analysis (FCA) considers attributes as a non-ordered set. This is appropriate when the data set is not structured. When an attribute taxonomy exists, existing techniques produce a completed context with all attributes deduced from the taxonomy. Usual algorithms can then be applied on the completed context for finding frequent concepts, but the results systematically contain redundant information. This article describes an algorithm which allows the frequent concepts of a formal context with taxonomy to be computed. It works on a non-completed context and uses the taxonomy information when needed. The results avoid the redundancy problem with equivalent performance.
    [bibtex-key = CFRD2006] [bibtex-entry]


  8. Sébastien Ferré. Negation, Opposition, and Possibility in Logical Concept Analysis.. In Rokia Missaoui and Jürg Schmid, editors, Int. Conf. Formal Concept Analysis, LNCS 3874, pages 130-145, 2006. Springer. Keyword(s): logic, concept analysis, epistemic logic, all i know. [bibtex-key = Fer2006b] [bibtex-entry]


  9. Yoann Padioleau, Benjamin Sigonneau, and Olivier Ridoux. LISFS: a logical information system as a file system (demo). In Leon J. Osterweil, H. Dieter Rombach, and Mary Lou Soffa, editors, Int. Conf. Software Engineering, pages 803-806, 2006. ACM. Keyword(s): demo. [bibtex-key = LISFS2006] [bibtex-entry]


  10. Eric Poupard, Denis Bechet, and Annie Foret. Categorial Grammar Acquisition from a French Treebank. In Actes de la Conférence d'APprentissage 2006 (CAP'06), 2006. Note: (Poster). [bibtex-key = Foret06a] [bibtex-entry]


  11. Benjamin Sigonneau and Olivier Ridoux. Software Engineering Applications of Logic File System -- Application to Automated Multi-Criteria Indexation of Software Components. In ICSE Work. Mining Software Repositories, 2006. [bibtex-key = SigRid2006a] [bibtex-entry]


  12. Alexandre Vautier, Marie-Odile Cordier, Mireille Ducassé, and René Quiniou. Agrégation d'alarmes faiblement structurées. In Actes de l'atelier ``Fouille de données temporelles'' associé aux 6es Journées Francophones ``Extraction et de Gestion des Connaissances'', Janvier 2006. [WWW]
    Abstract:
    La contribution principale de ce document est une approche plaçant l'opérateur au coeur de l'analyse de journaux d'alarmes faiblement structurées en lui permettant d'utiliser ce qu'il sait, même si ses connaissances sont partielles, et sans le submerger d'informations. Des motifs temporels structurés sont extraits par agrégation d'alarmes généralisées et corrélation se basant sur la date des alarmes et sur la similarité d'attributs autres que la date. L'approche est appliquée aux alarmes produites par un concentrateur VPN (Virtual Private Network). Une étude de cas montre comment 5000 d'alarmes peuvent être regroupées en 50 motifs.
    [bibtex-key = vautier06] [bibtex-entry]


  13. Sébastien Ferré and Olivier Ridoux. Logic Functors: A Toolbox of Components for Building Customized and Embeddable Logics. Research Report RR-5871, Irisa, March 2006. [WWW] Keyword(s): logic, components, modules and functors, theorem provers, type checking, application developpement.
    Abstract:
    Logic Functors form a framework for specifying new logics, and deriving automatically theorem provers and consistency/completeness diagnoses. Atomic functors are logics for manipulating symbols and concrete domains, while other functors are logic transformers that may add connectives or recursive structures, or may alter the semantics of a logic. The semantic structure of the framework is model theoretic as opposed to the verifunctional style often used in classical logic. This comes close to the semantics of description logics, and we show indeed that the logic~${\cal ALC}$ can be rebuilt using logic functors. This offers the immediate advantage that variants of~${\cal ALC}$ can be explored and implemented almost for free. This report comes with extensive appendices describing in detail a toolbox of logic functors (definitions, algorithms, theorems, and proofs).
    [bibtex-key = FerRid2006a] [bibtex-entry]


  14. Alexandre Vautier, Marie-Odile Cordier, Mireille Ducassé, and René Quiniou. Visualization of Internet Flow Records. Publication Interne 1828, IRISA, november 2006. Note: CURAR deliverable - CRE France Telecom R & D. [bibtex-key = creFT2006] [bibtex-entry]


2005
  1. Yoann Padioleau. Logic File System, un système de fichier basé sur la logique. Thèse d'université, Université de Rennes 1, February 2005. Note: Supervised by O. Ridoux. Keyword(s): file system, logical information system, navigation, file. [bibtex-key = Pad2005] [bibtex-entry]


  2. Sébastien Ferré and R. D. King. A dichotomic search algorithm for mining and learning in domain-specific logics. Fundamenta Informaticae -- Special Issue on Advances in Mining Graphs, Trees and Sequences, 66(1-2):1-32, 2005. [PDF] Keyword(s): machine learning, logic, concept analysis, data-mining, logic functors.
    Abstract:
    Many application domains make use of specific data structures such as sequences and graphs to represent knowledge. These data structures are ill-fitted to the standard representations used in machine learning and data-mining algorithms: propositional representations are not expressive enough, and first order ones are not efficient enough. In order to efficiently represent and reason on these data structures, and the complex patterns that are related to them, we use domain-specific logics. We show these logics can be built by the composition of logical components that model elementary data structures. The standard strategies of top-down and bottom-up search are ill-suited to some of these logics, and lack flexibility. We therefore introduce a dichotomic search strategy, that is analogous to a dichotomic search in an ordered array. We prove this provides more flexibility in the search, while retaining completeness and non-redundancy. We present a novel algorithm for learning using domain specific logics and dichotomic search, and analyse its complexity. We also describe two applications which illustrates the search for motifs in sequences; where these motifs have arbitrary length and length-constrained gaps. In the first application sequences represent the trains of the East-West challenge; in the second application they represent the secondary structure of Yeast proteins for the discrimination of their biological functions.
    [bibtex-key = FerKin2004b] [bibtex-entry]


  3. Denis Bechet, Alexandre Dikovsky, and Annie Foret. Dependency Structure Grammars. In Proceedings of the LACL 2005 Conference : Logical Aspects of Computational Linguistics, LNCS(LNAI) 3492, pages 18-34, 2005. springer. [PDF]
    Abstract:
    In this paper, we define Dependency Structure Grammars (DSG), which are rewriting rule grammars generating sentences together with their dependency structures, are more expressive than CF-grammars and non-equivalent to mildly context-sensitive grammars. We show that DSG are weakly equivalent to Categorial Dependency Grammars (CDG) recently introduced in [6,3]. In particular, these dependency grammars naturally express long distance dependencies and enjoy good mathematical properties.
    [bibtex-key = Foret05b] [bibtex-entry]


  4. Denis Bechet and Annie Foret. On Rigid NL Lambek Grammars Inference from Generalized Functor-Argument Data. In FGMOL'05, the tenth conference on Formal Grammar and the ninnth on the Mathematics of Language, Edinburgh, Scotland, 2005. Keyword(s): grammatical inference, categorial grammars, non- associative Lambek calculus, learning from positive examples, model of Gold.
    Abstract:
    This paper is concerned with the inference of categorial grammars, a context-free grammar formalism in the field of computational linguistics. A recent result has shown that whereas they are not learnable from strings in the model of Gold, rigid and k-valued non-asso ciative Lamb ek grammars are still learnable from generalized functor-argument structured sentences. We fo cus here on the algorithmic part of this result and provide an algo- rithm that can b e seen as an extension of Buszkowski, Penn and Kanazawa's contributions for classical categorial grammars.
    [bibtex-key = For05c] [bibtex-entry]


  5. Denis Bechet and Annie Foret. k-Valued Non-Associative Lambek Grammars (without Product) Form a Strict Hierarchy of Languages. In Proceedings of the LACL 2005 Conference: Logical Aspects of Computational Linguistics, LNCS(LNAI) 3492, pages 1-17, 2005. springer. [PDF]
    Abstract:
    The notion of k-valued categorial grammars where a word is associated to at most k types is often used in the field of lexicalized grammars as a fruitful constraint for obtaining several properties like the existence of learning algorithms. This principle is relevant only when the classes of k-valued grammars correspond to a real hierarchy of languages. This paper establishes the relevance of this notion for two related grammatical systems. In the first part, the classes of k-valued non-associative Lambek (NL) grammars without product is proved to define a strict hierarchy of languages. The second part introduces the notion of generalized functor argument for non-associative Lambek ($NL_{\emptyset}$) calculus without product but allowing empty antecedent and establishes also that the classes of k-valued ($NL_{\emptyset}$) grammars without product form a strict hierarchy of languages
    [bibtex-key = Foret05a] [bibtex-entry]


  6. T. Denmat, M. Ducassé, and O. Ridoux. Data mining and cross-checking of execution traces. A re-interpretation of Jones, Harrold and Stasko test information visualization. In T. Ellman and A. Zisman, editors, Proceedings of the 20th IEEE/ACM International Conference on Automated Software Engineering, November 2005. ACM Press. Note: See RR-5661 for a long version of this article. Keyword(s): Software Engineering, Debugging, Artificial Intelligence, Learning, Knowledge acquisition.
    Abstract:
    The current trend in debugging and testing is to cross-check information collected during several executions. Jones et al., for example, propose to use the instruction coverage of passing and failing runs in order to visualize suspicious statements. This seems promising but lacks a formal justification. In this paper, we show that the method of Jones et al. can be re-interpreted as a data mining procedure. More particularly, they define an indicator which characterizes association rules between data. With this formal framework we are able to explain intrinsic limitations of the above indicator.
    [bibtex-key = denmat05b] [bibtex-entry]


  7. T. Denmat, A. Gotlieb, and Mireille Ducassé. Proving or Disproving Likely Invariants with Constraint Reasoning. In A. Serebrenik, editor, Proceedings of the 15th Workshop on Logic-based Method for Programming Environments, Sitges, SPAIN, October 2005. Note: Satelite event of International Conference on Logic Programming (ICLP'2005). Published in Computer Research Repository cs.SE/0508108. [WWW] Keyword(s): Software Engineering, Testing and Debugging, Program verification, Constraint and logic languages.
    Abstract:
    A program invariant is a property that holds for every execution of the program. Recent work suggest to infer likely-only invariants, via dynamic analysis. A likely invariant is a property that holds for some executions but is not guaranteed to hold for all executions. In this paper, we present work in progress addressing the challenging problem of automatically verifying that likely invariants are actual invariants. We propose a constraint-based reasoning approach that is able, unlike other approaches, to both prove or disprove likely invariants. In the latter case, our approach provides counter-examples. We illustrate the approach on a motivating example where automatically generated likely invariants are verified.
    [bibtex-key = DGD05] [bibtex-entry]


  8. Sébastien Ferré, Olivier Ridoux, and Benjamin Sigonneau. Arbitrary Relations in Formal Concept Analysis and Logical Information Systems. In ICCS, LNCS 3596, pages 166-180, 2005. Springer. Keyword(s): logical concept analysis, relation, logical information system, navigation.
    Abstract:
    A logical view of formal concept analysis considers attributes of a formal context as unary predicates. In a first part, we propose an augmented definition that handles {\em binary relations} between objects. A Galois connection is defined on augmented contexts. It represents concept inheritance as usual, but also relations between concepts. As usual, labeling operators are also defined. In particular, concepts and relations are visible and labeled in a single structure. In a second part, we show how relations can be used for navigating in an augmented concept lattice. This part augments the theory of Logical Information Systems. An implementation is sketched, and first experimental results are presented.
    [bibtex-key = FerRidSig2005] [bibtex-entry]


  9. L. Langevine and M. Ducassé. A Tracer Driver for Hybrid Execution Analyses. In Proceedings of the 6th Automated Debugging Symposium, September 2005. ACM Press. Note: See RR-5611 for a longer version of this article. [WWW] Keyword(s): Software Engineering, Debugging, Monitors, Tracing, Programming Environments.
    Abstract:
    Tracers provide users with useful information about program executions. In this paper we propose a ``tracer driver'', from a single tracer, it provides a powerful front-end for multiple dynamic analysis tools while limiting the overhead of the trace generation. The tracer driver can be used both synchronously and asynchronously. The relevant execution events are specified by flexible event patterns and a large variety of trace data can be given either systematically or ``on demand''. The proposed tracer driver has been designed and experimented in the context of constraint logic programming, within GNU-Prolog. Its principles are, however, independent of the traced programming language. Experimental measures show that the flexibility and power of the described architecture are also the basis of reasonable performances.
    [bibtex-key = langevine05] [bibtex-entry]


  10. L. Langevine and M. Ducassé. A Tracer Driver for Versatile Dynamic Analyses of Constraint Logic Programs. In A. Serebrenik, editor, Proceedings of the 15th Workshop on Logic-based Method for Programming Environments, Sitges, SPAIN, October 2005. Note: Satelite event of International Conference on Logic Programming (ICLP'2005). Published in Computer Research Repository cs.SE/0508105. [WWW] Keyword(s): Software Engineering, Debugging, Monitors, Tracing, Programming Environments.
    Abstract:
    Programs with constraints are hard to debug. In this paper, we describe a general architecture to help develop new debugging tools for constraint programming. The possible tools are fed by a single general-purpose tracer. A tracer-driver is used to adapt the actual content of the trace, according to the needs of the tool. This enables the tools and the tracer to communicate in a client-server scheme. Each tool describes its needs of execution data thanks to event patterns. The tracer driver scrutinizes the execution according to these event patterns and sends only the data that are relevant to the connected tools. Experimental measures show that this approach leads to good performance in the context of constraint logic programming, where a large variety of tools exists and the trace is potentially huge.
    [bibtex-key = langevine05b] [bibtex-entry]


  11. Yoann Padioleau and Olivier Ridoux. A Parts-of-File File System. In USENIX Annual Technical Conference, General Track (Short Paper), 2005. [WWW]
    Abstract:
    The Parts-of-file File System (PofFS) allows read-write accesses to different views of a given file or set of files in order to help the user separate and manipulate different concerns. The set of files is considered as a mount point from which views can be selected as read-write files via directories. Paths are formulas mentioning properties of a desired view. Each directory contain a file (the view) which contains the parts of the mounted files that satisfy the properties. This service is offered generically at the file system level, and a plug-in interface permits that file formats, or application-specific details are handled by user-defined operators. Special plug-ins called transducers can be defined for automatically attaching properties to parts of files. Performances are encouraging; files of 100 000 lines are handled efficiently.
    [bibtex-key = padioleau05] [bibtex-entry]


  12. Yoann Padioleau, Benjamin Sigonneau, Olivier Ridoux, and Sébastien Ferré. LISFS: a Logical Information System as a File System. In Véronique Benzaken, editor, Bases de données avancées, pages 393-398, October 2005. Université de Rennes 1. [WWW] [PDF]
    Abstract:
    We present Logical Information Systems (LIS). A LIS can be viewed as a schema-less database whose objects are described by logical formulas. Objects are automatically organized according to their logical description, and logical formulas can be used for representing both queries and navigation links. The key feature of a LIS is that it answers a query with a set of navigation links expressed in the same logic as the query. As navigation links are dynamically computed from any query, and can be used as query increments, it follows that querying and navigation steps can be combined in any order. We then present LISFS, a file-system implementation of a LIS, where objects are files or parts of files. This has the benefit to make LIS features available right now to existing applications. This implementation can easily be extended and specialized through a plug-in mechanism. Finally, we present some applications in the field of personal databases (e.g., music, images, emails), and demonstrate that building specialized interfaces for visualizing databases can be done easily through LISFS navigation.
    [bibtex-key = lisfs-bda05] [bibtex-entry]


  13. T. Denmat, M. Ducassé, and O. Ridoux. Data Mining and Cross-checking of Execution Traces. A re-interpretation of Jones, Harrold and Stasko test information visualization (Long version). Research Report RR-5661, INRIA, August 2005. Note: Also Publication Interne IRISA PI-1743. [WWW] [bibtex-key = denmat05c] [bibtex-entry]


  14. L. Langevine and M. Ducassé. A Tracer Driver to Enable Concurrent Dynamic Analyses. Research Report RR-5611, INRIA, June 2005. [WWW] [bibtex-key = langevine05c] [bibtex-entry]


2004
  1. M. Ducassé, U. Nilsson, and D. Seipel, editors. Proceedings of the First International Workshop on Teaching Logic Programming: TeachLP, http://www.ep.liu.se/ecp/012/, September 2004. Linköping Electronic Conference Proceedings, Issue No. 12. Note: ISSN 1650-3686 (print), 1650-3740 (www). [bibtex-key = ducasse04b] [bibtex-entry]


  2. L. Langevine, P. Deransart, and M. Ducassé. A Generic Trace Schema for the Portability of CP(FD) Debugging Tools. In J. Vancza, K. Apt, F. Fages, F. Rossi, and P. Szeredi, editors, Recent advances in Constraint Programming, pages 171-195. Springer-Verlag, Lecture Notes in Artificial Intelligence 3010, 2004. [bibtex-key = langevine04] [bibtex-entry]


  3. Hervé Debar, B. Morin, Frédéric Cuppens, Fabien Autrel, Ludovic Mé, Bernard Vivinis, Salem Benferhat, M. Ducassé, and Rodolphe Ortalo. Corrélation d'alertes en détection d'intrusions. Technique et Science Informatiques, 23(3):323-358, 2004. [bibtex-key = debar04] [bibtex-entry]


  4. Sébastien Ferré and Olivier Ridoux. An Introduction to Logical Information Systems. Information Processing & Management, 40(3):383-419, 2004. Keyword(s): information systems, information search, retrieval, query formulation, representation languages, deduction, theorem proving.
    Abstract:
    Logical information systems (LIS) use logic in a uniform way to describe their contents, to query it, to navigate through it, to analyze it, and to maintain it. They can be given an abstract specification that does not depend on the choice of a particular logic, and concrete instances can be obtained by instantiating this specification with a particular logic. In fact, a logic plays in a LIS the role of a schema in databases. We present the principles of LIS, the constraints they impose on the expression of logics, and hints for their effective implementation.
    [bibtex-key = FerRid2004] [bibtex-entry]


  5. Jeanne Villaneau, Olivier Ridoux, and Jean-Yves Antoine. LOGUS : un système formel de compréhension de l'oral spontané. RIA --- Revue d'Intelligence Artificielle, (à paraître), 2004. [bibtex-key = VRA-RIA2004] [bibtex-entry]


  6. D. Bechet, A. Dikovsky, A. Foret, and E. Moreau. On learning discontinuous dependencies from positive data. In Proceedings of the Formal Grammar Conference (FG 2004), August 2004. [bibtex-key = Foret04b] [bibtex-entry]


  7. Denis Bechet and Annie Foret. On Intermediate Structures for Non-Associative Lambek Grammars and Learnability. In Proceedings of the CG 2004 Conference : Categorial Grammars An efficient tool for Natural Language Processing, June 2004. [bibtex-key = Foret04a] [bibtex-entry]


  8. D. Bechet, A. Foret, and I. Tellier. Learnability of Pregroup Grammars. In Proceedings of the International Conference of grammatical Inference (ICGI 2004), LNAI 3264, October 2004. [bibtex-key = Foret04c] [bibtex-entry]


  9. Sébastien Ferré and R. D. King. BLID: an Application of Logical Information Systems to Bioinformatics. In P. Eklund, editor, Int. Conf. Formal Concept Analysis, LNCS 2961, pages 47-54, 2004. Springer. Keyword(s): logical concept analysis, information system, bioinformatics.
    Abstract:
    BLID (Bio-Logical Intelligent Database) is a bioinformatic system designed to help biologists extract new knowledge from raw genome data by providing high-level facilities for both data browsing and analysis. We describe BLIDrsquos novel data browsing system which is based on the idea of Logical Information Systems. This enables combined querying and navigation of data in BLID (extracted from public bioinformatic repositories). The browsing language is a logic especially designed for bioinformatics. It currently includes sequence motifs, taxonomies, and macromolecule structures, and it is designed to be easily extensible, as it is composed of reusable components. Navigation is tightly combined with this logic, and assists users in browsing a genome through a form of human-computer dialog.
    [bibtex-key = FerKin2004a] [bibtex-entry]


  10. L. Langevine and M. Ducassé. A tracer driver to enable debugging, monitoring and visualization of CLP executions from a single tracer. In B. Demoen and V. Lifschtitz, editors, Proceedings of the International Conference on Logic Programming, pages 462-463, September 2004. Lecture Notes in Computer Science 3132, Springer-Verlag. Note: Poster. [bibtex-key = langevine04c] [bibtex-entry]


  11. L. Langevine and M. Ducassé. Un pilote de traceur pour la PLC. Déboguer, auditer et visualiser une exécution avec un même traceur. In F. Mesnard, editor, Actes des Journées Francophones de Programmation en Logique avec Contraintes, pages 19-36, Juin 2004. HERMES Science Publications. [bibtex-key = langevine04b] [bibtex-entry]


  12. Benjamin Sigonneau and Olivier Ridoux. Indexation multiple et automatisée de composants logiciels orientés objet. In AFADL --- Approches Formelles dans l'Assistance au Développement de Logiciels, 2004. [bibtex-key = RS-AFADL2004] [bibtex-entry]


  13. E. Tombini, H. Debar, L. Mé, and M. Ducassé. A serial combination of anomaly and misuse IDSes applied to HTTP traffic. In D. Thomsen and C. Schuba, editors, Proceedings of the Annual Computer Security Applications Conference, December 2004. [bibtex-key = tombini04] [bibtex-entry]


  14. Mireille Ducassé and Benjamin Sigonneau. Building efficient tools to query execution traces. Rapport de Recherche RR-5280, INRIA, July 2004. Note: Also Publication Interne IRISA 1638. [bibtex-key = ducasse04] [bibtex-entry]


  15. J.-P. Pouzol, S. Benferhat, H. Debar, M. Ducassé, E. Fayol, S. Gombault, J. Goubault-Larrecq, Y. Lavictoire, L. Mé, L. Noé, J. Olivain, E. Totel, and B. Vivinis. Rapport de synthèse sur la création de sondes de détection d'intrusions. Livrable du projet RNTL DICO, Juillet 2004. Note: 121 pages. [bibtex-key = dico-sp4-04] [bibtex-entry]


2003
  1. M. Ducassé, editor. Actes des Journées Francophones de Programmation en Logique avec Contraintes, June 2003. Revue des Sciences et Technologies de l'Information, Hors série/JFPLC 2003, HERMES Science Publications. [bibtex-key = ducasse03] [bibtex-entry]


  2. Denis Bechet and Annie Foret. Apprentissage des grammaires de Lambek rigides et d'arité bornée pour le traitement automatique des langues. In Actes de la Conférence d'APprentissage 2003 (CAP'2003), June 2003. [bibtex-key = Bechet03g] [bibtex-entry]


  3. Denis Bechet and Annie Foret. Remarques et perspectives sur les langages de prégroupe d'ordre 1/2. In ATALA, editor, Actes de la conférence Traitement Automatique des Langues Naturelles (TALN'2003), June 2003. Note: (Poster). [bibtex-key = Bechet03e] [bibtex-entry]


  4. Denis Bechet and Annie Foret. k-Valued Non-Associative Lambek Grammars are Learnable from Function-Argument Structures. In Proceedings of the 10th Workshop on Logic, Language, Information and Computation (WoLLIC'2003), volume 85, Electronic Notes in Theoretical Computer Science, July 2003. [bibtex-key = Bechet03f] [bibtex-entry]


  5. Denis Bechet and Annie Foret. k-valued Non-Associative Lambek Categorial Grammars are not Learnable from Strings. In ACL, editor, Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics (ACL 2003), July 2003. [bibtex-key = Bechet03d] [bibtex-entry]


  6. P. Deransart, L. Langevine, and M. Ducassé. A Generic Trace Schema for the Portability of CP(FD) Debugging Tools. In Proceedings of the ERCIM workshop on Constraint and Logic Programming, 2003. [bibtex-key = deransart03] [bibtex-entry]


  7. M. Ducassé, L. Langevine, and P. Deransart. Rigorous design of tracers: an experiment for constraint logic programming. In M. Ronsse, editor, Proceedings of the Fifth International Workshop on Automated Debugging, September 2003. CoRR cs.SE/0309027. [bibtex-key = ducasse03b] [bibtex-entry]


  8. Sébastien Ferré. The Use of Associative Concepts for Fast Incremental Concept Formation in Sparse Contexts. In B. Ganter and A. de Moor, editors, Using Conceptual Structures -- Contributions to ICCS 2003, 2003. Shaker Verlag. [PDF] Keyword(s): concept analysis, associative concept, lattice, algorithm, sparse context. [bibtex-key = Fer2003] [bibtex-entry]


  9. Annie Foret. On the Computation of Joins for Non Associative Lambek Categorial Grammars.. In Proceedings of the 17th International Workshop on Unification Valencia, Spain, June 8-9, (UNIF'03), 2003. [bibtex-key = For03u] [bibtex-entry]


  10. L. Langevine, M. Ducassé, and P. Deransart. A Propagation Tracer for Gnu-Prolog: from Formal Definition to Efficient Implementation. In C. Palamidessi, editor, Proceedings of the 19th Int. Conf. in Logic Programming, December 2003. Springer-Verlag, Lecture Notes in Computer Science. [bibtex-key = langevine03] [bibtex-entry]


  11. Yoann Padioleau and Olivier Ridoux. A Logic File System. In Proc. USENIX Annual Technical Conference, 2003. [bibtex-key = yp-or-usenix2003] [bibtex-entry]


  12. Yoann Padioleau and Olivier Ridoux. Présentation du «Parts-of-file File System». In Proc. Conférence Française sur les Systèmes d'exploitation, 2003. [bibtex-key = yp-or-cfse2003] [bibtex-entry]


  13. Olivier Ridoux. Logic Information Systems for Logic Programmers. In C. Palamidessi, editor, Proc. Int. Conf. Logic Programming, 2003. Springer-Verlag, LNCS. [bibtex-key = or-iclp2003] [bibtex-entry]


  14. Jean Goubault-Larrecq, Stéphane Demri, M. Ducassé, Ludovic Mé, Julien Olivain, Claudine Picaronny, Jean-Philippe Pouzol, Eric Totel, and Bernard Vivinis. Algorithmes de détection et langages de signatures. Livrable 3.3, Projet RNTL DICO, http://dico.netsecuresoftware.com//index.htm, Octobre 2003. [bibtex-key = dico-sp3-03] [bibtex-entry]


  15. Yoann Padioleau and Olivier Ridoux. A Logic File System. Research Report 4656, INRIA, 2003. [WWW] Keyword(s): file system, information retrieval, applied logic.
    Abstract:
    We present the new paradigm of logic file systems, its implementation, and first experimental results. It offers in an integrated way navigation and classification, the possibility of expressive queries, ease of use, and possible heterogeneity of data. This paradigm is object-centered. It associates logical descriptions to objects, and logical relations between descriptions serve as a basis for navigation and querying. We compare logic file systems with the hierarchical, boolean, and data-base paradigms. We present briefly the rôle of logic in logic file systems, and in more details the implementation issues of a particular logic file system that uses a simple logic.
    [bibtex-key = yp-or-lisfs-inria2003] [bibtex-entry]


  16. Yoann Padioleau and Olivier Ridoux. The Parts-of-file File System. Rapport de recherche 4783, INRIA, 2003. [bibtex-key = yp-or-poffs-inria2003] [bibtex-entry]


  17. P. Deransart, L. Langevine, and M. Ducassé. Debugging Constraint problems with Portable Tools. Demonstration presented at the 13th Workshop on Logic Programming Environments, December 2003. Keyword(s): demo. [bibtex-key = deransart03b] [bibtex-entry]


2002
  1. M. Ducassé, editor. Journal of Automated Software Engineering 9(1), special issue on automated debugging, janvier 2002. Kluwer. [bibtex-key = ducasse02] [bibtex-entry]


  2. Sébastien Ferré. Systèmes d'information logiques : un paradigme logico-contextuel pour interroger, naviguer et apprendre. Thèse d'université, Université de Rennes 1, October 2002. Note: Accessible en ligne à l'adresse http://www.irisa.fr/bibli/publi/theses/theses02.html.[WWW] Keyword(s): système d'information, treillis (théorie des), analyse de concepts, logique, recherche de l'information, navigateurs (logiciels), apprentissage automatique.
    Abstract:
    Les deux principaux paradigmes de recherche d'information, la navigation et l'interrogation, sont souvent dissociés. Les systèmes hiérarchiques offrent une structure de navigation figée qui ne convient pas à toutes les utilisations ; ce qu'ils compensent par des outils de recherche. Ceux-ci, fondés sur l'interrogation, sont plus souples mais sont plus difficiles à utiliser pour les non-initiés et rendent délicat le contrôle du volume des réponses. Il apparaît donc comme nécessaire de combiner étroitement navigation et interrogation. Pour réaliser cette combinaison, nous nous fondons sur l'Analyse de concepts (AC) qui permet de construire automatiquement, à partir d'une description des objets, une structure de navigation appelée «treillis de concepts», où les concepts jouent à la fois le rôle de répertoire et de requête. Comme dans l'AC les descriptions se limitent à des ensembles d'attributs, nous avons généralisé l'AC pour les remplacer par des formules d'une logique arbitraire. Ceci nous semble important pour traiter des applications diverses. Les Systèmes d'information logiques (SIL) se définissent donc par la combinaison navigation/interrogation, l'emploi de la logique (descriptions, requêtes et liens de navigation) et la généricité. Sur cette base, nous avons développé plusieurs mécanismes pour faciliter l'expression et la découverte de connaissances. Les connaissances d'un domaine peut être exprimées par une terminologie. Un dialogue homme-machine, fondé sur le treillis de concepts, permet de retrouver des objets (navigation) et de découvrir des régularités entre les objets (extraction de connaissances). Un mécanisme d'apprentissage offre une assistance à la classification des objets. Enfin, un prototype a été développé pour d'expérimenter ces mécanismes. Il est générique dans le sens où il ne dépend pas de la logique employée. Ces logiques peuvent être assemblés à l'aide d'un jeu de composants logique, que nous avons constitué.
    [bibtex-key = Fer2002a] [bibtex-entry]


  3. E. Jahier and M. Ducassé. Generic Program Monitoring by Trace Analysis. Theory and Practice of Logic Programming, 2(4-5), September 2002. [bibtex-key = jahier02] [bibtex-entry]


  4. Soazig Bars, Sébastien Ferré, and Olivier Ridoux. Logic Functors for Types as Search Keys. In Int. Workshop on Isomorphisms of Types, 2002. [WWW] [bibtex-key = BarFerRid2002] [bibtex-entry]


  5. P. Deransart, L. Langevine, and M. Ducassé. A Generic Trace Model for Finite Domain Solvers. In B. O'Sullivan, editor, Proc. of the International Workshop on User-Interaction in Constraint Satisfaction, September 2002. Cornell University. [bibtex-key = deransart02] [bibtex-entry]


  6. M. Ducassé and L. Langevine. Analyse automatisée de traces d'exécution de programmes CLP(FD). In M. Rueher, editor, Actes des Journées Francophones de Programmation en Logique avec Contraintes, pages 119-134, Mai 2002. HERMES science publications. [bibtex-key = ducasse02c] [bibtex-entry]


  7. M. Ducassé and L. Langevine. Automated analysis of CLP(FD) program execution traces. In P. Stuckey, editor, Proceedings of the International Conference on Logic Programming, July 2002. Lecture Notes in Computer Science 2401, Springer-Verlag. Note: Poster. Extended version available at http://www.irisa.fr/LIS/ducasse/. [bibtex-key = ducasse02d] [bibtex-entry]


  8. Sébastien Ferré and Olivier Ridoux. The Use of Associative Concepts in the Incremental Building of a Logical Context. In G. Angelova U. Priss, D. Corbett, editor, Int. Conf. Conceptual Structures, LNCS 2393, pages 299-313, 2002. Springer. Keyword(s): concept analysis, logic, information systems, learning, classification, context.
    Abstract:
    A formal context associates to objects a description that combines automatically extracted properties (intrinsic) and manually assigned ones (extrinsic). The extrinsic properties are expressed by users according to intentions that are often subjective and changing, and determine the classification and retrieval of objects. So, we find it important to assist users in this task through the automatic suggestion of extrinsic properties to be assigned and even the discovery of rules to automate these assignements. The principle is to learn from the description of existing objects the extrinsic description of a new object. Because of the changing nature of users' intentions, the assistance given in the incremental building of a logical context must be interactive. We present formal principles, and an application to the classification of email messages.
    [bibtex-key = FerRid2002] [bibtex-entry]


  9. Annie Foret and Yannick Le Nir. Lambek rigid grammars are not learnable from strings. In COLING'2002, 19th International Conference on Computational Linguistics, Taipei, Taiwan, 2002. [bibtex-key = For02b] [bibtex-entry]


  10. Annie Foret and Yannick Le Nir. Les grammaires de Lambek rigides ne sont pas apprenables à partir des chaines. In CAp'2002, Conférence Francophone d'Apprentissage, Orléans, France, 2002. Presses Universitaires de Grenoble. [bibtex-key = For02a] [bibtex-entry]


  11. Annie Foret and Yannick Le Nir. On limit points for some variants of rigid Lambek Grammars. In ICGI'2002, the 6th International Colloquium on Grammatical Inference, number 2484 of Lecture Notes in Artificial Intelligence, Amsterdam, the Netherlands, 2002. Springer-Verlag. [bibtex-key = For02c] [bibtex-entry]


  12. L. Langevine, P. Deransart, M. Ducassé, and E. Jahier. Prototypage de traceurs CLP(FD). In M. Rueher, editor, Actes des Journées Francophones de Programmation en Logique avec Contraintes, pages 135-150, Mai 2002. HERMES science publications. [bibtex-key = langevine02] [bibtex-entry]


  13. B. Morin, L. Mé, H. Debar, and M. Ducassé. M2D2 : A Formal Data Model for IDS Alert Correlation. In A. Wespi et al., editor, Recent Advances in Intrusion Detection, pages 97-104, October 2002. Springer-Verlag, Lecture Notes in Computer Science 2516. [bibtex-key = morin02] [bibtex-entry]


  14. Jean-Philippe Pouzol and M. Ducassé. Formal specification of intrusion signatures and detection rules. In S. Schneider, editor, Proc. of 15th IEEE Computer Security Foudations Workshop, pages 64-76, 2002. IEEE Press. [bibtex-key = pouzol02] [bibtex-entry]


  15. J. Villaneau, J.-Y. Antoine, and Olivier Ridoux. LOGUS : un système formel de compréhension du français parlé spontané - présentation et évaluation. In TALN'02: Traitement Automatique de la Langue Naturelle, 2002. [bibtex-key = var-taln2002] [bibtex-entry]


  16. Sébastien Ferré. Incremental Concept Formation made More Efficient by the Use of Associative Concepts. Research Report RR-4569, Inria, Institut National de Recherche en Informatique et en Automatique, October 2002. [WWW] Keyword(s): context, concept lattice, Galois lattice, incremental algorithm, complexity.
    Abstract:
    Formal Concept Analysis (FCA) is interested in the formation of concept lattices from binary relations between objects and attributes, a.k.a. contexts. Many algorithms have been proposed to generate the set of all concepts, and also the edges of the lattice between these concepts. We develop the principle and the code of a new algorithm combining two existing ones, Godin's and Bordat's algorithms. Then, we show by both a theoretical and practical study that it is the most efficient algorithm for sparse contexts, which are usually found in real applications.
    [bibtex-key = Fer2002] [bibtex-entry]


  17. Sébastien Ferré and Olivier Ridoux. Introduction to Logical Information Systems. Research Report RR-4540, Inria, Institut National de Recherche en Informatique et en Automatique, September 2002. [WWW] Keyword(s): information system, information search and retrieval, query formulation, representation languages, deduction and theorem proving.
    Abstract:
    Logical Information Systems (LIS) use logic in a uniform way to describe their contents, to query it, to navigate through it, to analyze it, and to maintain it. They can be given an abstract specification that does not depend on the choice of a particular logic, and concrete instances can be obtained by instantiating this specification with a particular logic. In fact, a logic plays in a LIS the role of a schema in data-bases. We present the principles of logical information systems, the constraints they impose on the expression of logics, and hints for their effective implementation.
    [bibtex-key = FerRid2002a] [bibtex-entry]


  18. Sébastien Ferré and Olivier Ridoux. Logic Functors : a Framework for Developing Embeddable Customized Logics. Technical report RR-4457, Inria, Institut National de Recherche en Informatique et en Automatique, May 2002. [WWW] Keyword(s): applied logic, software component.
    Abstract:
    Logic-based applications often use customized logics which are composed of several logics. These customized logics are also often embedded as a black-box in an application. So, implementing them requires the specification of a well-defined interface with common operations such as a parser, a printer, and a theorem prover. In order to be able to compose these logic, one must also define composition laws, and prove their properties. We present the principles of logic functors and their compositions for constructing logics that are ad-hoc, but sound. An important issue is how the operations of different sublogics inter-operate. We propose a formalization of the logic functors, their semantics, implementations, proof-theoretic properties, and their composition.
    [bibtex-key = FerRid2002b] [bibtex-entry]


2001
  1. M. Ducassé and E. Jahier. Efficient Automated Trace Analysis: Examples with Morphine. Electronic Notes in Theoretical Computer Science, 55(2), 2001. Note: K. Havelund and G. Rosu (Eds), proceedings of the first Workshop on Runtime Verification. [bibtex-key = ducasse01] [bibtex-entry]


  2. O. Ridoux and Patrice Boizumault. Typed Static Analysis: Application to the Groundness Analysis of Typed Prolog. Journal of Functional and Logic Programming, 2001(4), July 2001. [bibtex-key = rb-jflp2001] [bibtex-entry]


  3. Sébastien Ferré. Complete and Incomplete Knowledge in Logical Information Systems. In Salem Benferhat and Philippe Besnard, editors, Symbolic and Quantitative Approaches to Reasoning with Uncertainty, LNCS 2143, pages 782-791, 2001. Springer. Keyword(s): modal logic, All I know, complete and incomplete knowledge, information system.
    Abstract:
    We present a generalization of logic All I Know by presenting it as an extension of standard modal logics. We study how this logic can be used to represent complete and incomplete knowledge in Logical Information Systems. In these information systems, a knowledge base is a collection of objects (e.g., files, bibliographical items) described in the same logic as used for expressing queries. We show that usual All I Know (transitive and euclidean accessibility relation) is convenient for representing complete knowledge, but not for incomplete knowledge. For this, we use \emph{serial} All I Know (serial accessibility relation).
    [bibtex-key = Fer2001] [bibtex-entry]


  4. Sébastien Ferré and Olivier Ridoux. A Framework for Developing Embeddable Customized Logics. In A. Pettorossi, editor, Int. Work. Logic-based Program Synthesis and Transformation, LNCS 2372, pages 191-215, 2001. Springer. [WWW] Keyword(s): logic, composition, theorem prover, logic-based systems.
    Abstract:
    Logic-based applications often use customized logics which are composed of several logics. These customized logics are also often embedded as a black-box in an application. Their implementation requires the specification of a well-defined interface with common operations such as a parser, a printer, and a theorem prover. In order to be able to compose these logics, one must also define composition laws, and prove their properties. We present the principles of logic functors and their compositions for constructing customized logics. An important issue is how the operations of different sublogics inter-operate. We propose a formalization of the logic functors, their semantics, implementations, and their composition.
    [bibtex-key = FerRid2001b] [bibtex-entry]


  5. Sébastien Ferré and Olivier Ridoux. Searching for Objects and Properties with Logical Concept Analysis. In Harry S. Delugach and Gerd Stumme, editors, International Conference on Conceptual Structures, LNCS 2120, pages 187-201, 2001. Springer. Keyword(s): concept analysis, navigation, knowledge discovery, logical information system.
    Abstract:
    Logical Concept Analysis is Formal Concept Analysis where logical formulas replace sets of attributes. We define a Logical Information System that combines navigation and querying for searching for objects. Places and queries are unified as formal concepts represented by logical formulas. Answers can be both extensional (objects belonging to a concept) and intensional (formulas refining a concept). Thus, all facets of navigation are formalized in terms of Logical Concept Analysis. We show that the definition of being a refinement of some concept is a specific case of Knowledge Discovery in a formal context. It can be generalized to recover more classical KD~operations like machine-learning through the computation of necessary or sufficient properties (modulo some confidence), or data-mining through association rules.
    [bibtex-key = FerRid2001] [bibtex-entry]


  6. Annie Foret. Conjoinability and unification in Lambek categorial grammars. In New Perspectives in Logic and Formal Linguisitics, Proceedings Vth ROMA Workshop, Roma, 2001. Bulzoni Editore. [bibtex-key = For01a] [bibtex-entry]


  7. Annie Foret. Mixing deduction and substitution in Lambek categorial grammars, some investigations. In LACL'01, 4th International Conference on Logical Aspects of Computational Linguistics, number 2099 of Lecture Notes in Artificial Intelligence, Le Croisic, France, 2001. Springer-Verlag. [bibtex-key = For01b] [bibtex-entry]


  8. Annie Foret. The emptiness of intersection problem for k-valued categorial grammars (classical and Lambek) is undecidable. In FGMOL'01, the sixth conference on Formal Grammar and the seventh on the Mathematics of Language, Helsinki, Finland, 2001. [bibtex-key = For01c] [bibtex-entry]


  9. E. Jahier, M. Ducassé, and O. Ridoux. Specifying Prolog Trace Models with a Continuation Semantics. In K.-K. Lau, editor, Logic Based Program Synthesis and Transformation, 2001. Springer-Verlag, Lecture Notes in Computer Science 2042. [bibtex-key = jahier01b] [bibtex-entry]


  10. L. Langevine, P. Deransart, M. Ducassé, and E. Jahier. Prototyping clp(fd) tracers: a trace model and an experimental validation environment. In T. Kusalik, editor, Proceedings of the 11th Workshop on Logic Programming Environments, 2001. Computer Research Repository, CS.PL/0111043. [bibtex-key = langevine01] [bibtex-entry]


  11. J.-P. Pouzol and M. Ducassé. From Declarative Signatures to Misuse IDS. In W. Lee, L. Mé, and A. Wespi, editors, Recent Advances in Intrusion Detection, Proceedings of the 4th International Symposium, pages 1-21, 2001. Springer-Verlag, Lecture Notes in Computer Science 2212. [bibtex-key = pouzol01] [bibtex-entry]


  12. J. Villaneau, J.-Y. Antoine, and Olivier Ridoux. Combining Syntax and Pragmatic Knowledge for the Understanding of Spontaneous Spoken Sentences. In Ph. de Groote, G. Morill, and Ch. Retoré, editors, 4th Int. Conf. Logical Aspects of Computational Linguistics, 2001. [bibtex-key = var-lacl2001] [bibtex-entry]


  13. M. Ducassé and L. Rozé. Revisiting the ``Traffic lights'' B case study. Publication Interne 1424, IRISA, November 2001. [bibtex-key = ducasse01b] [bibtex-entry]


2000
  1. M. Ducassé, editor. Proceedings of the 4th International Workshop on Automated Debugging (AADEBUG2000), Munich, August 2000. COrr. Note: Refereed proceedings to appear in the COmputer Research Repository (CORR). [WWW] [bibtex-key = ducasse2000c] [bibtex-entry]


  2. M. Ducassé and J. Noyé. Tracing Prolog programs by source instrumentation is efficient enough. Elsevier Journal of Logic Programming,, 43(2):157-172, May 2000. [WWW] Keyword(s): Debugging, tracing, source to source transformation, benchmarking, Prolog..
    Abstract:
    Tracing by automatic program source instrumentation has major advantages over compiled code instrumentation: it is more portable, it benefits from many compiler optimizations, it produces traces in terms of the original program, and it can be tailored to specific debugging needs. The usual argument in favor of compiled code instrumentation is its supposed efficiency. We have compared the performance of two operational low-level Prolog tracers with source instrumentation. We have executed classical Prolog benchmark programs, collecting trace information without displaying it. On average, collecting trace information by program instrumentation is about as fast as using a low-level tracer in one case, and only twice slower in the other. This is a minor penalty to pay, compared to the advantages of the approach. To our knowledge, this is the first time that a quantitative comparison of both approaches is made for any programming language.
    [bibtex-key = ducasse2000] [bibtex-entry]


  3. V. Abily and M. Ducassé. Benchmarking a distributed intrusion detection system based on ASAX: Preliminary results. In H. Debar, editor, RAID 2000 (Recent Advances on Intrusion Detection), 2000. Note: Refereed extended abtract. [WWW] [bibtex-key = abily2000] [bibtex-entry]


  4. M. Ducassé and L. Rozé. Proof obligations of the B formal method: Local proofs ensure global consistency. In A. Bossi, editor, LOgic-based Program Synthesis and TRansformation, pages 11-30, 2000. Springer-Verlag, Lecture Notes in Computer Science, 1817. [bibtex-key = ducasse2000b] [bibtex-entry]


  5. Sébastien Ferré and Olivier Ridoux. A File System Based on Concept Analysis. In Yehoshua Sagiv, editor, International Conference on Rules and Objects in Databases, number 1861 of Lecture Notes in Computer Science, pages 1033-1047, July 2000. Springer. Keyword(s): concept analysis, logic, information system, file system.
    Abstract:
    We present the design of a file system whose organization is based on Concept Analysis ``à la Wille-Ganter''. The aim is to combine querying and navigation facilities in one formalism. The file system is supposed to offer a standard interface but the interpretation of common notions like directories is new. The contents of a file system is interpreted as a Formal Context, directories as Formal Concepts, and the sub-directory relation as Formal Concepts inclusion. We present an organization that allows for an efficient implementation of such a Conceptual File System.
    [bibtex-key = FerRid2000a] [bibtex-entry]


  6. Sébastien Ferré and Olivier Ridoux. A Logical Generalization of Formal Concept Analysis. In Guy Mineau and Bernhard Ganter, editors, International Conference on Conceptual Structures, number 1867 of Lecture Notes in Computer Science, pages 371-384, August 2000. Springer. Keyword(s): concept analysis, logic, context, information system.
    Abstract:
    We propose a generalization of Formal Concept Analysis (FCA) in which sets of attributes are replaced by expressions of an almost arbitrary logic. We prove that all FCA can be reconstructed on this basis. We show that from any logic that is used in place of sets of attributes can be derived a contextualized logic that takes into account the formal context and that is isomorphic to the concept lattice. We then justify the generalization of FCA compared with existing extensions and in the perspective of its application to information systems.
    [bibtex-key = FerRid2000b] [bibtex-entry]


  7. E. Jahier, M. Ducassé, and O. Ridoux. Specifying Prolog Trace Models with a Continuation Semantics. In K.-K. Lau, editor, Proc. of LOgic-based Program Synthesis and TRansformation, London, July 2000. Note: Technical Report Report Series, Department of Computer Science, University of Manchester, ISSN 1361-6161. Report number UMCS-00-6-1.[WWW] [bibtex-key = jahier00] [bibtex-entry]


  8. E. Jahier, M. Ducassé, and O. Ridoux. Spécification de modèles de traces de programmes Prolog à l'aide d'une sémantique par continuation. In Touraivane, editor, Actes des Journées francophones de Programmation Logique et par Contraintes, Marseille, 2000. Hermès. [bibtex-key = jahier00b] [bibtex-entry]


  9. D. Le Métayer, V.-A. Nicolas, and Olivier Ridoux. Verification by testing for recursive program schemes. In A. Bossi, editor, LOgic-based Program Synthesis and TRansformation, 2000. Springer-Verlag, Lecture Notes in Computer Science, 1817. [bibtex-key = LOPSTR99] [bibtex-entry]


  10. J.-P. Pouzol and M. Ducassé. Handling Generic Intrusion Signatures is not Trivial. In H. Debar, editor, RAID 2000 (Recent Advances on Intrusion Detection), 2000. Note: Refereed extended abtract. [WWW] [bibtex-key = pouzol2000] [bibtex-entry]


  11. Sébastien Ferré and Olivier Ridoux. A File System Based on Concept Analysis. Technical Report RR-3942, Inria, Institut National de Recherche en Informatique et en Automatique, avril 2000. [WWW] [bibtex-key = FerRid2000c] [bibtex-entry]


1999
  1. M. Ducassé, A. Kusalik, and G. Puebla, editors. Proceedings of the 10th Workshop on Logic Programming Environments, http://www.cs.usask.ca/projects/envlop/WLPE/10WLPE/, novembre 1999. [bibtex-key = ducasse99g] [bibtex-entry]


  2. S. Mallet. Explications dans les bases de données déductives : Associer trace et sémantique. PhD thesis, Insa de Rennes, Irisa, novembre 1999. [bibtex-key = smallet99d] [bibtex-entry]


  3. M. Ducassé. Abstract views of Prolog executions with Opium. In P. Brna, B. du Boulay, and H. Pain, editors, Learning to Build and Comprehend Complex Information Structures: Prolog as a Case Study, Cognitive Science and Technology. Ablex, 1999. [WWW] Keyword(s): Software engineering, Programming environment, Automated debugging, Trace abstraction mechanisms, Debugging language, Program behavior understanding, Prolog Debugging tool.
    Abstract:
    Opium is a system for analysing and debugging Prolog programs. Its kernel comprises an execution tracer and a programming language with a set of primitives for trace and source analysis. In this chapter we show the power of Opium for supporting abstract views of Prolog executions. Abstract views give high-level points of view about executions. They filter out irrelevant details; they restructure the remaining information; and they compact it so that the amount of information given at each step has a reasonable size. The examples of abstract views given in the following are a goal execution profile, some data abstractions, an instantiation profile, a failure analysis, a loop analysis, and a kind of explanation for an expert system written in Prolog.
    [bibtex-key = ducasse99b] [bibtex-entry]


  4. C. Belleannée, P. Brisset, and Olivier Ridoux. A Pragmatic Reconstruction of $\lambda$Prolog. J. Logic Programming, 41(1):67-102, 1999. [bibtex-key = BEL99] [bibtex-entry]


  5. M. Ducassé. Opium: An extendable trace analyser for Prolog. The Journal of Logic programming, 1999. Note: Special issue on Synthesis, Transformation and Analysis of Logic Programs, A. Bossi and Y. Deville (eds), Also Rapport de recherche INRIA RR-3257 and Publication Interne IRISA PI-1127. [WWW] Keyword(s): Software Engineering, Automated Debugging, Trace Query Language, Program Execution Analysis, Abstract Views of Program Executions, Prolog.
    Abstract:
    Traces of program executions are a helpful source of information for program debugging. They, however, give a picture of program executions at such a low level that users often have difficulties to interpret the information. Opium, our extendable trace analyzer, is connected to a ``standard'' Prolog tracer. Opium is programmable and extendable. It provides a trace query language and abstract views of executions. Users can therefore examine program executions at the levels of abstraction which suit them. Opium has shown its capabilities to build abstract tracers and automated debugging facilities. This article describes in depth the trace query mechanism, from the model to its implementation. Characteristic examples are detailed. Extensions written so far on top of the trace query mechanism are listed. Two recent extensions are presented: the abstract tracers for the LO (Linear Objects) and the CHR (Constraint Handling Rules) languages. These two extensions were specified and implemented within a few days. They show how to use Opium for real applications.
    [bibtex-key = ducasse99] [bibtex-entry]


  6. M. Ducassé. An introduction to the B formal method. In A.-L. Bossi, editor, Proceedings of the 9th International Workshop on LOgic-based Program Synthesis and TRansformation, pages 23-30, September 1999. Universita' Ca' Foscari di Venezia. Note: Technical report CS-99-16, Slides. [bibtex-key = ducasse99d] [bibtex-entry]


  7. M. Ducassé. Coca: An automated Debugger for C. In Proceedings of the 21st International Conference on Software Engineering, pages 504-513, May 1999. ACM Press. [WWW] Keyword(s): Software engineering, Programming environment, Automated debugging, Trace query mechanism, Debugging language, Program behavior understanding, C Debugging tool..
    Abstract:
    We present Coca, an automated debugger for C, where the breakpoint mechanism is based on events related to language constructs. Events have semantics whereas source lines used by most debuggers do not have any. A trace is a sequence of events. It can be seen as an ordered relation in a database. Users can specify precisely which events they want to see by specifying values for event attributes. At each event, visible variables can be queried. The trace query language is Prolog with a handful of primitives. The trace query mechanism searches through the execution traces using both control flow and data whereas debuggers usually search according to either control flow or data. As opposed to fully ``relational'' debuggers which use plain database querying mechanisms, Coca trace querying mechanism does not require any storage. The analysis is done on the fly, synchronously with the traced execution. Coca is therefore more powerful than ``source line'' debuggers and more efficient than relational debuggers.
    Annotation:
    Also RR-3489
    [bibtex-key = ducasse99c] [bibtex-entry]


  8. E. Jahier and M. Ducassé. A generic approach to monitor program executions. In D. De Schreye, editor, Proceedings of the International Conference on Logic Programming, November 1999. MIT Press. [WWW] Keyword(s): Monitoring, Trace analysis, Flexibility, Logic programming, Mercury.
    Abstract:
    Monitoring requires to gather data about executions. The monitoring functionalities currently available are built on top of ad hoc instrumentations. Most of them are implemented at low-level; in any case they require an in-depth knowledge of the system to instrument. The best people to implement these instrumentations are generally the implementors of the compiler. They, however, cannot decide which data to gather. Indeed, hundreds of variants can be useful and only end-users know what they want. In this article, we propose a primitive which enables users to easily specify what to monitor. It is built on top of the tracer of the Mercury compiler. We illustrate how to use this primitive on two different kinds of monitoring. Firstly, we implement monitors that collect various kinds of statistics; each of them is well-known, the novelty is that users can get exactly the variants they need. Secondly, we define two notions of test coverage for logic programs and show how to measure coverage rates with our primitive. To our knowledge no definition of test coverage exist for logic programming so far. Each example is only a few lines of Mercury. Measurements show that the performance of the primitive on the above examples is acceptable for an execution of several millions of trace events. Our primitive, although simple, lays the foundation for a generic and powerful monitoring environment.
    [bibtex-key = jahier99d] [bibtex-entry]


  9. E. Jahier and M. Ducassé. Un traceur d'exécutions de programmes ne sert pas qu'au débogage. In F. Fages, editor, Actes des Journées francophones de Programmation Logique et par Contraintes, Lyon, juin 1999. Hermès. [WWW] Keyword(s): Dynamic analysis, Trace analysis, Monitoring, Measure of test coverage, Logic programming, Mercury.
    Abstract:
    In this paper we show that a tracer with a trace analyser can be used to achieve more than debugging. We first illustrate how to compute coverage ratios for test cases. We also give 4 examples to monitor the behavior of programs. Thus, instead of building ad hoc instrumentations, which is currently the case for such tools, one can use a uniform environment which allows a synergy between the tools to take place. As a matter of fact, while studying the test coverage measurement we enriched the trace information, to the benefit of the other tools. Moreover, ad hoc instrumentations require an in depth knowledge of the system to instrument, either at low level or by source to source transformation. Even it is not technically difficult, it always requires a significant programming effort. On the opposite, with our approach, the instrumentation is generic. It is done once for all, and the specific analyses can be relatively simple. The examples of this article consist of less than a dozen of Prolog lines each.
    [bibtex-key = jahier99b] [bibtex-entry]


  10. E. Jahier, M. Ducassé, and O. Ridoux. Specifying trace models with a continuation semantics. In M. Ducassé, A. Kusalik, L. Naish, and G. Puebla, editors, Proc. of ICLP'99 Workshop on Logic Programming Environments, 1999. Note: LPE'99. [WWW] Keyword(s): trace models, continuation semantics, specification, validation, Logic programming.
    Abstract:
    In this article, we have give a formal specification of Byrd's box model and we show how this specification can be extended to specify richer trace models. We have also sho how these specifications can be executed by a direct translation into lambda-Prolog, leading to a Prolog interpreter that performs execution traces. This interpreter can be used both to experiment various trace models and to validate the different event specifications. Hence we have a formal framework to specify and prototype trace models.
    [bibtex-key = jahier99e] [bibtex-entry]


  11. S. Mallet and M. Ducassé. Generating deductive database explanations. In D. De Schreye, editor, Proceedings of the International Conference on Logic Programming, November 1999. MIT Press. [WWW] Keyword(s): deductive databases, debugging, trace, operational semantics, multi-SLD-AL, meta-interpreter, substitution set, instrumentation.
    Abstract:
    Existing explanation systems for deductive databases show forests of proof trees. Although proof trees are often useful, they are only one possible interesting representation. We argue that an explanation system for deductive databases must be able to generate explanations at several levels of abstraction. One possible and well known technique to achieve this flexibility is to instrument meta-interpreters. It is, however, not often used because of its inefficiency. On the other hand, deductive databases often generate intermediate information stored in the physical database. This information can be considered as a low-level trace giving a faithful picture of what has happened at the relational level. The deductive reasoning is lost but can be very easily recovered by a meta-interpreter. In this article we describe a technique to generate explanations by integrating a relational trace and an instrumented meta-interpreter. The expensive aspects of meta-interpretation are reduced by the use of the trace which avoids many costly calculations. The flexibility of meta-interpretation is preserved, as illustrated by the generation of three different kinds of explanations: a box-oriented trace, a multi-SLD-AL tree and abstract AND trees. This technique enables powerful explanation systems to be implemented with very few modifications of the deductive database mechanism itself.
    [bibtex-key = sm-iclp99] [bibtex-entry]


  12. S. Mallet and M. Ducassé. Myrtle: A set-oriented meta-interpreter driven by a ``relational'' trace for deductive database debugging. In P. Flener, editor, LOgic-based Program Synthesis and TRansformation, pages 328-330, 1999. Springer-Verlag, LNCS 1559. Note: Résumé, version complète en RR-3598 INRIA. [bibtex-key = smallet99b] [bibtex-entry]


  13. D. Le Métayer, V.-A. Nicolas, and O. Ridoux. Verification by testing for recursive program schemes. In LOPSTR'99 (International Workshop on Logic Program Synthesis and Transformation), 1999. Springer-Verlag, LNCS. [WWW] Keyword(s): Software engineering, program verification, white-box testing, automated test data generation, program analysis, program schemes.
    Abstract:
    In this paper, we explore the testing-verification relationship with the objective of mechanizing the generation of test data. We consider program classes defined as recursive program schemes and we show that complete and finite test data sets can be associated with such classes, that is to say that these test data sets allow us to distinguish every two different functions in these schemes. This technique is applied to the verification of simple properties of programs.
    [bibtex-key = dlm-van-or-lopstr99] [bibtex-entry]


  14. Olivier Ridoux, Patrice Boizumault, and Frédéric Malésieux. Typed static analysis: application to groundness analysis of LambdaProlog and Prolog. In A. Middeldorp and T. Sato, editors, Proc. of the Int. Symp. on Functional and Logic Programming, 1999. Springer. [bibtex-key = RID99] [bibtex-entry]


  15. Sébastien Ferré and Olivier Ridoux. Une généralisation logique de l'analyse de concepts logique. Technical Report RR-3820, Inria, Institut National de Recherche en Informatique et en Automatique, December 1999. [WWW] Keyword(s): concept analysis, concept lattice, logic, context, information systems, querying, browsing.
    Abstract:
    Nous proposons une généralisation de l'analyse de concepts formels (ACF) dans laquelle les ensembles d'attributs sont remplacés par des expressions d'une logique presque arbitraire. Nous prouvons que toute l'ACF peut être reconstruite sur cette base. Nous montrons qu'à partir de toute logique utilisée à la place des ensembles d'attributs, on peut dériver une logique contextualisée qui prend en compte le contexte formel et qui est isomorphe au treillis de concepts. Nous comparons ensuite la généralisation de l'ACF aux extensions qui y ont déjà été apportées. Enfin, nous présentons nos perspectives d'application aux systèmes d'information.
    Annotation:
    FCA,logic,SI
    [bibtex-key = FerRid1999] [bibtex-entry]


  16. S. Mallet and M. Ducassé. Myrtle: A set-oriented meta-interpreter driven by a ``relational'' trace for deductive database debugging. Research Report RR-3598, INRIA, January 1999. [WWW] Keyword(s): deductive databases, debugging, trace, multi-SLD-AL, meta-interpreter, substitution set, instrumentation.
    Abstract:
    Deductive databases manage large quantities of data and, in general, in a set-oriented way. The existing systems of explanation for deductive databases do not take these constraints into account. We propose a tracing technique which consists of integrating a "relational" trace and an instrumented meta-interpreter using substitution sets. The relational trace efficiently gives precise information about data extraction from the relational database. The meta-interpreter manages substitution sets and gives explanation on the deduction. The expensive aspects of meta-interpretation are reduced by the use of the trace which avoids many calculations. The flexibility of meta-interpretation is preserved. It allows different profiles of trace to be easily produced.
    [bibtex-key = sm-rr99] [bibtex-entry]


  17. M. Ducassé and E. Jahier. An automated debugger for Mercury - Opium-M 0.1 User and reference manuals. mai 1999. Note: RT-231 INRIA (aussi PI-1234 IRISA). [bibtex-key = ducasse99e] [bibtex-entry]


  18. E. Jahier and M. Ducassé. Opium-M 0.1 User and Reference Manuals. IRISA, Rennes, March 1999. [WWW] Keyword(s): Logic programming, Mercury, Trace analyser, Trace query language, Automated debugging, User manual, Reference manual.
    Abstract:
    This document gathers the user manual and the reference manual of Opium-M, an analyser of execution traces of Mercury Programs. Opium-M is an adaptation to Mercury of Opium a trace analyser for Prolog. Mercury is a new logic programming language. Its type, mode and determinism declarations enable codes to be generated that is at the same time more efficient and more reliable than with current logic programming languages. The deterministic parts of Mercury programs are as efficient as their C counterparts. Moreover, numerous mistakes are detected at compilation time. However, our industrial partner experience shows that the fewer remaining mistakes, the harder they are to be diagnosed. A high-level debugging tool was thus necessary. Program execution traces given by traditional debuggers provide programmers with useful pieces of information. However, using them requires to analyse by hand huge amounts of information. Opium-M is connected to the traditional tracer of Mercury, it allows execution trace analyses to be automated. It provides a relational trace query language based on Prolog which enables users to specify precisely what they want to see in the trace. Opium-M, then, automatically filters out information irrelevant for the users.
    [bibtex-key = jahier99] [bibtex-entry]


1998
  1. D. Le Métayer, V.-A. Nicolas, and O. Ridoux. Programs, Properties, and Data: Exploring the Software Development Trilogy. IEEE Software, 15(6):75-81, November/December 1998. [WWW] Keyword(s): Software engineering, testing, verification, program analysis, program learning.
    Abstract:
    Software development usually involves a collection of properties, programs and data as input or output documents. Putting these three kinds of documents at the vertices of a triangle, one sees that all three sides of the triangle have been exploited in formal methods, and that they have often been used in both directions. However, richer combinations have seldom been envisaged, and formal methods often amount to a strict orientation of the figure by imposing functional dependencies (e.g.,~infering test cases from specifications). Moreover, undecidability problems arise when properties are expressed in full predicate logic (or similar formalisms) or programs are written in Turing-equivalent programming languages. We advocate that (1) formal methods should provide more flexible ways to exploit the developer's knowledge and offer a variety of possibilities to construct programs, properties and test data and (2) it is worth restricting the power of logic formalisms and programming languages for the benefit of mechanization. We go one step in this direction, and present a formal method for generating test cases that combines techniques from abstract interpretation ({\em program -> property}) and testing ({\em program+property -> test data}), and takes inspiration from automated learning (test generation via a {\em testing bias}). The crucial property of the test suites generated this way is that they are robust with respect to a test objective formalized as a property. In other words, if a program passes the test suite, then it is guaranteed to satisfy the property. As this process leads to decision problems in very restricted formalisms, it can be fully mechanized.
    [bibtex-key = dlm-van-or-IEEEsoftware98] [bibtex-entry]


  2. M. Ducassé. Teaching B at a Technical University is Possible and Rewarding. In H. Habrias and S. E. Dunn, editors, B'98, Proceedings of the Educational Session, avril 1998. Association de Pilotage des Conférences B, Nantes. Note: ISBN: 2-9512461-0-2. [WWW] Keyword(s): B formal method, teaching.
    Abstract:
    In January 1994, to replace a highly unpopular denotational semantics course, I undertook to set up a course on the B method at the INSA of Rennes (Institut National des Sciences Appliquées), at a Bac+4 level. I had almost no previous knowledge of formal methods. I had, however, programmed much in Prolog and felt the need for a strong programming discipline, supported if possible by methods and tools. The experience is, in my opinion, successful. The students do learn much during the course, find interesting placements where their competence is appreciated and every occurrence of the course teaches me something. In the article, I first list reasons to start the experience. I then discuss the pedagogical objectives of the course. The contents of the course is given and an assessment is made.
    [bibtex-key = ducasse98c] [bibtex-entry]


  3. M. Ducassé and J. Noyé. Tracing Prolog Programs by Source Instrumentation is Efficient Enough. In K. Sagonas, editor, IJCSLP'98 Post-conference workshop on Implementation Technologies for Programming Languages based on Logic., June 1998. [WWW] Keyword(s): Debugging, tracing, source to source transformation, benchmarking, Prolog.
    Abstract:
    Tracing by automatic program source instrumentation has major advantages over compiled code instrumentation: it is more portable, it benefits from many compiler optimizations, it produces traces in terms of the original program, and it can be tailored to specific debugging needs. The usual argument in favor of compiled code instrumentation is its supposed efficiency. We have compared the performance of two operational low-level Prolog tracers with source instrumentation. We have executed classical Prolog benchmark programs, collecting trace information without displaying it. On average, collecting trace information by program instrumentation is about as fast as using a low-level tracer in one case, and only twice slower in the other. This is a minor penalty to pay, compared to the advantages of the approach. To our knowledge, this is the first time that a quantitative comparison of both approaches is made for any programming language.
    [bibtex-key = ducasse98b] [bibtex-entry]


  4. S. Mallet and M. Ducassé. Pilotage d'un méta-interprète ensembliste par une trace ``relationnelle'' pour le débogage de bases de données déductives. In O. Ridoux, editor, Journées francophones de Programmation Logique et programmation par Contraintes, Nantes, pages 151-165, mai 1998. JFPLC'98, Hermes. [WWW] Keyword(s): bases de données déductives, débogage, trace, sémantique opérationnelle, multi-SLD-AL, méta-interprète, ensembles de substitutions, instrumentation.
    Abstract:
    Le développement des bases de données déductives nécessite des outils, en particulier pour le débogage. Les bases de données déductives gèrent des quantités importantes de données et, en général, de manière ensembliste. Les systèmes d'explication existants pour les bases de données déductives ne prennent pas en compte ces contraintes. Nous proposons une technique de traçage qui consiste à intégrer une trace ``relationnelle'' avec un méta-interprète instrumenté utilisant des ensembles de substitutions. La trace relationnelle donne, de manière efficace, de l'information précise sur l'extraction de données de la base relationnelle. Le méta-interprète ensembliste gère des ensembles de substitutions et donne des explications sur la déduction. Les aspects coûteux de la méta-interprétation sont réduits par l'utilisation de la trace qui évite beaucoup de calculs. La flexibilité de la méta-interprétation est conservée. Elle permet de produire facilement des traces de profils différents.
    [bibtex-key = mallet98a] [bibtex-entry]


  5. M. Ducassé. Coca: A Debugger for C Based on Fine Grained Control Flow and Data Events. Technical report IRISA PI 1202 or INRIA RR-3489, INRIA, septembre 1998. [WWW] Keyword(s): Software engineering, Programming environment, Automated debugging, Trace query mechanism, Debugging language, Program behavior understanding, C Debugging tool.
    Abstract:
    We present Coca, an automated debugger for C, where the breakpoint mechanism is based on events related to language constructs. Events have semantics whereas source lines used by most debuggers do not have any. A trace is a sequence of events. It can be seen as an ordered relation in a database. Users can specify precisely which events they want to see by specifying values for event attributes. At each event, visible variables can be queried. The trace query language is Prolog with a handful of primitives. The trace query mechanism searches through the execution traces using both control flow and data whereas debuggers usually search according to either control flow or data. As opposed to fully ``relational'' debuggers which use plain database querying mechanisms, Coca trace querying mechanism does not require any storage. The analysis is done on the fly, synchronously with the traced execution. Coca is therefore more powerful than ``source line'' debuggers and more efficient than relational debuggers.
    [bibtex-key = ducasse98] [bibtex-entry]


1997
  1. S. Mallet and M. Ducassé. An Informal Presentation of DDB Trees: A Basis for Deductive Database Explanations. In Ulrich Geske, editor, DDLP'97, Fifth International Workshop on Deductive Databases and Logic Programming, July 1997. GMD-Studien. Keyword(s): debugging, explanations, deductive databases, logic programming. [bibtex-key = mallet97b] [bibtex-entry]


  2. S. Mallet and M. Ducassé. DDB trees: a basis for deductive database explanations. In Mariam Kamkar, editor, AADEBUG'97,Third International Workshop on Automated Debugging, Linköping, Sweden, pages 87-102, May 1997. [WWW] Keyword(s): debugging, explanations, deductive databases, logic programming.
    Abstract:
    The power of deductive systems in general is that programs express what should be done and not how it should be done. Nevertheless, deductive systems need debugging and explanation facilities. Indeed, their operational semantics is less abstract than the declarative semantics of the programs. If users have to understand all the low level details of the operational semantics much of the benefits of using a deductive system is lost. Existing explanation systems for deductive databases produce proof trees to be shown to users. Although useful, proof trees give a fragmented view of query evaluations, and users face a, most of the time large, forest of proof trees. We propose a new data structure, called the DDB tree, which merges the information of a proof tree forest into one concise tree. A DDB tree gives a global picture of a query evaluation in a dramatically reduced structure with no loss of information. DDB trees can be shown to users or can be analyzed further by an explanation system.
    [bibtex-key = mallet97a] [bibtex-entry]


1996
  1. M. Ducassé and J. Noyé. Tracing Prolog without a tracer. In N. Fuchs and U. Geske, editors, Proceedings of the poster session at JICSLP'96, pages 223-232, September 1996. GMD- Forschungszentrum Informationstechnik GMBH, GMD-STUDIEN Nr.296, ISBN3-88457-296-2. Note: One page abstract also appears in Proc. of the JICSLP'96, MIT Press, ISBN 0-262-63173-3.
    Abstract:
    Tracing by automatic program source instrumentation has major advantages over compiled code instrumentation: it is cheaper to develop and more portable, it benefits from most compiler optimizations, it produces traces in terms of the original program, and it can be tailored to specific debugging needs. The usual argument in favor of compiled code instrumentation is its supposed efficiency. Tolmach and Appel 1 designed and implemented a tracer for Standard ML based on automatic program source instrumentation. The resulting code runs only 3 times slower than optimized code. They conjectured that a low-level tracer would run at about the same speed. However they had no reasonable low-level tracer at hand to actually compare their results with. We have performed such a comparison in the context of Prolog, using the ECRC ECLiPSe environment. The built-in low-level tracer of ECLiPSe is, at present, one of the most interesting tracers for Prolog. We have compared it with an instrumentation based on O'Keefe's "advice" utility 2 , made compatible with the ECLiPSe tracer. We traced "standard" Prolog benchmark programs 3 with both tracing techniques and measured the resulting CPU times. On average the performances of both implementations are equivalent: tracing Prolog programs by program instrumentation is no slower than using a low-level tracer. To our knowledge, this is the first time that a quantitative comparison of both approaches is made. Another contribution is that our source instrumentation is more complete than O'Keefe's advice package. In particular, it deals with built-in predicates, and allows predicates to be skipped/unskipped. ---------------------------------------- 1 A. Tolmach and A.W. Appel. A debugger for Standard ML. Journal of Functional Programming, 5(2):155-200, April 1995. 2 The "advice" utility is part of the DEC10 Prolog library, available by anonymous ftp from the AIAI of the University of Edinburg (aiai.edinburgh.ac.uk). 3 P. Van Roy and A. M. Despain. High-performance logic programming with the Aquarius Prolog compiler. Computer, 25(1):54-68, January 1992.
    [bibtex-key = dn96] [bibtex-entry]


  2. P. Louvet and O. Ridoux. Parametric Polymorphism for Typed Prolog and $\lambda$Prolog. In 8th Int. Symp. Programming Languages Implementation and Logic Programming, volume 1140 of LNCS, Aachen, Germany, pages 47-61, 1996. [WWW] Keyword(s): Logic programming, typing, polymorphism, second-order lambda-calculus..
    Abstract:
    Typed Prolog and LambdaProlog are logic programming languages with a strict typing discipline which is based on simple types with variables. These variables are interpreted as denoting generic polymorphism. Experiments show that this discipline does not handle properly common logic programming practices used in Prolog. For instance, the usual transformation for computing the Clark completion of a Prolog program does not work well with some typed programs. We observe that the head-condition is at the heart of these problems, and conclude that it should be enforced. We propose a second-order scheme which is compatible with usual practices. In this scheme, type variables denote parametric polymorphism. It allows quantifying types and terms, passing type and term parameters to goals and terms, and to express type guards for selecting goals. We give its syntax and deduction rules, and propose a solution to keep the concrete notation of programs close to the usual one.
    [bibtex-key = louvet:parametric:plilp:96] [bibtex-entry]


  3. O. Ridoux. Engineering Transformations of Attributed Grammars in $\lambda$Prolog. In M. Maher, editor, Joint Int. Conf. and Symp. Logic Programming, pages 244-258, 1996. MIT Press. [WWW] Keyword(s): Syntax-directed translation, grammar transformations, logic grammars, DCG, LambdaProlog..
    Abstract:
    An abstract representation for grammar rules that permits an easy implementation of several attributed grammar transformations is presented. It clearly separates the actions that contribute to evaluating attribute values from the circulation of these values, and it makes it easy to combine the representations of several rules in order to build the representation of new rules. This abstract form applies well to such transforms as elimination of left-recursion, elimination of empty derivation, unfolding and factorization. Finally, the technique is applied to DCGs and a LambdaProlog implementation of the abstract form and of the transforms is described.
    [bibtex-key = ridoux:engineering:jicslp:96] [bibtex-entry]


  4. S. Schoenig and M. Ducassé. A Backward Slicing Algorithm for Prolog. In R. Cousot and D.A. Schmidt, editors, Static Analysis Symposium, Aachen, pages 317-331, September 1996. Springer-Verlag, LNCS 1145. [WWW]
    Abstract:
    Slicing is a program analysis technique originally developed by Weiser for imperative languages. Weiser showed that slicing is a natural tool for debugging, but it has other numerous applications (program integration, program optimization, etc.) In this article we describe a backward slicing algorithm for Prolog which produces executable slices. The proposed algorithm is applicable at least to pure Prolog extended by some simple built-in predicates that handle the explicit unification =/2 and arithmetic. To our knowledge, this algorithm is the first one to be proposed for Prolog. Because of the indeterminism and lack of explicit control flow of Prolog, existing algorithms cannot be trivially adapted. The two main contributions of this paper are a general definition of slicing adapted to Prolog and a slicing algorithm that produces executable programs.
    [bibtex-key = sm96] [bibtex-entry]


  5. S. Schoenig and M. Ducassé. Slicing pour programmes Prolog. In Actes des journées GDR programmation'96, Orléans, Novembre 1996. Université de Bordeaux I. [WWW]
    Abstract:
    Le slicing est une technique d'analyse de programme développée à l'origine par Weiser pour les langages impératifs. Weiser a montré que le slicing est un outil naturel de débogage, mais il a également de nombreuses autres applications (intégration de programmes, optimisation, etc.) Dans cet article, nous proposons une définition du slicing pour Prolog et un algorithme. Celui-ci est au moins applicable à Prolog pur étendu par quelques prédicats de base (=/2 et arithmétiques). À notre connaissance, cet algorithme est le premier à être proposé pour Prolog. Les spécificités de Prolog (indéterminisme et manque de flot de contr\^ole explicite), ne permettent pas d'adapter trivialement les algorithmes existants pour langages impératifs.
    [bibtex-key = sm96b] [bibtex-entry]


1995
  1. M. Ducassé, editor. Proceedings of the 2nd International Workshop on Automated and Algorithmic Debugging, Saint Malo, France, May 1995. IRISA, Campus de Beaulieu, F-35042 Rennes cedex. Note: See http://www.irisa.fr/EXTERNE/manifestations/AADEBUG95/. [bibtex-key = duc95b] [bibtex-entry]


  2. C. Belleannée, P. Brisset, and O. Ridoux. Une reconstruction pragmatique de $\lambda$Prolog. Technique et science informatiques, 14:1131-1164, 1995. [WWW] Keyword(s): Programmation logique, LambdaProlog, lambda-calcul, quantifications, types. logic programming, LambdaProlog, lambda-calculus, quantifications, types..
    Abstract:
    LambdaProlog est un langage de programmation logique dont les clauses et les termes généralisent ceux de Prolog. On peut se demander si toutes ces extensions sont nécessaires simultanément et si des langages intermédiaires intéressants ne pourraient pas être définis, au moins dans un but pédagogique. Nous répondons à cette question en montrant que des liens de nécessité conduisent à adopter toutes les extensions à partir de l'introduction du nouveau domaine de termes. De cette reconstruction découle une heuristique de programmation par induction sur les types qui est un guide commode pour utiliser LambdaProlog. LambdaProlog is a logic programming language in which clauses and terms are more general than in Prolog. One may wonder whether these extensions are simultaneously needed, and what are the useful subsets of LambdaProlog, at least for pedagogical purposes. We answer this question by exhibiting necessity links from the addition of the new term domain to the extension of the formula language. A handy heuristic for programming by induction on types can be derived from these links.
    [bibtex-key = belleannee:reconstruction:tsi:95] [bibtex-entry]


  3. S. Coupet-Grimal and O. Ridoux. On the use of Advanced Logic Programming Languages in Computational Linguistics. J. Logic Programming, 24(1&2):121-159, 1995. [WWW] Keyword(s): Logic programming, computational linguistics, LambdaProlog, Prolog II, lambda-terms, rational terms..
    Abstract:
    Computational Linguistics and Logic Programming have strong connections, but the former uses concepts that are absent from the most familiar implementations of the latter. We advocate that a Logic Programming language need not feature the Computational Linguistics concepts exactly, it must only provide a logical way of dealing with them. We focus on the manipulation of higher-order terms and the logical handling of context, and we show that the advanced features of Prolog~II and LambdaProlog are useful for dealing with these concepts. Higher-order terms are native in LambdaProlog, and Prolog~II's infinite trees provide a handy data-structure for manipulating them. The formula language of LambdaProlog can be transposed in the Logic Grammar realm to allow for a logical handling of context.
    [bibtex-key = coupet:use:jlp:95] [bibtex-entry]


  4. M. Ducassé. Automated Debugging Extensions of the Opium Trace Analyser. In Proceedings of the 2nd International Workshop on Automated and Algorithmic Debugging, Saint Malo, France, May 1995. IRISA, Campus de Beaulieu, F-35042 Rennes cedex. [WWW]
    Abstract:
    Traces of program executions tell how programs behave in given cases. They are a helpful source of information for automated debugging. Opium is an automated trace analyser for Prolog programs. It is programmable and extendable. It provides a trace query language and abstract views of executions as a basis for automated debugging. Opium has shown its capabilities to build abstract tracers and automated debugging facilities. This paper lists the extensions written so far, and describes two recent extensions: the abstract tracers for the LO (Linear Objects) language and for the CHR (Constraint Handling Rules) language.
    [bibtex-key = duc95] [bibtex-entry]


  5. O. Ridoux. Imagining CLP$(\Lambda,\equiv_{\alpha\beta})$. In A. Podelski, editor, Constraint Programming: Basics and Trends. Selected papers of the 22nd Spring School in Theoretical Computer Science. LNCS 910, Châtillon/Seine, France, pages 209-230, 1995. [WWW] Keyword(s): CLP, LambdaProlog, lambda-calculus..
    Abstract:
    We study under which conditions the domain of lambda-terms and the equality theory of the lambda-calculus form the basis of a usable constraint logic programming language (CLP). The conditions are that the equality theory must contain axiom {$\eta$}, and the formula language must depart from Horn clauses and accept universal quantifications and implications in goals. In short, CLP-lambda must be close to LambdaProlog.
    [bibtex-key = ridoux:imagining:sctcs:95] [bibtex-entry]


1994
  1. M. Ducassé and J. Noyé. Logic Programming Environments: Dynamic program analysis and debugging. The Journal of Logic Programming, 19/20:351-384, May/July 1994. Note: Anniversary issue: Ten years of Logic Programming. [WWW]
    Abstract:
    Programming environments are essential for the acceptance of programming languages. This survey emphasizes that program analysis, both static and dynamic, is the central issue of programming environments. Because their clean semantics makes powerful analysis possible, logic programming languages have an indisputable asset in the long term. This survey is focused on logic program analysis and debugging. The large number of references provided show that the field, though maybe scattered, is active. A unifying framework is given which separates environment tools into extraction, analysis, and visualization. It facilitates the analysis of existing tools and should give some guide lines to develop new ones. Achievements in logic programming are listed; some techniques developed for other languages are pointed out, and some trends for further research are drawn. Among the main achievements are algorithmic debugging, tracing for sequential Prolog, and abstract interpretation. The main missing techniques are slicing, test case generation, and program mutation. The perspectives we see are integration, evaluation and, above all, automated static and dynamic analysis.
    [bibtex-key = dn94] [bibtex-entry]


  2. P. Brisset and O. Ridoux. The Architecture of an Implementation of $\lambda$Prolog: Prolog/Mali. In ILPS'94 Workshop on Implementation Techniques for Logic Programming Languages, 1994. [WWW] [bibtex-key = brisset:architecture:ilpsw:94] [bibtex-entry]


1993
  1. P. Brisset and O. Ridoux. Continuations in $\lambda$Prolog. In D.S. Warren, editor, 10th Int. Conf. Logic Programming, pages 27-43, 1993. MIT Press. [WWW] Keyword(s): LambdaProlog, compilation, continuation, exception handling..
    Abstract:
    Continuations are well know in functional programming where they have been used to transform and compile programs. Some languages provide explicit manipulations of the continuation for the user: The user can catch and modify the current continuation. Continuations have also been used in the logic programming context to give a denotational semantics for Prolog, to generate Prolog compilers and to transform Prolog programs. In this paper, we propose to introduce new built-ins in a logic programming language to enable the user to explicitly replace the continuations. These built-ins allow the user to have a new control of the execution. We choose LambdaProlog because of its higher-order syntax and implications in the goals which are necessary for the definition and use of these built-ins. In order to define the built-ins, we extend to LambdaProlog the Prolog semantics based on continuations. Then, we show that an exception mechanism can be easily implemented using these new built-ins. The proposed semantics is also used to prove equivalence of goals changing the continuations.
    [bibtex-key = brisset:continuations:iclp:93] [bibtex-entry]


  2. M. Ducassé. A pragmatic survey of automated debugging. In P. Fritzson, editor, Proceedings of the First Workshop on Automated and Algorithmic Debugging, volume 749 of Lecture Notes in Computer Sciences, Linkoeping, Sweden, May 1993. Springer-Verlag. [WWW]
    Abstract:
    This article proposes a structuring view of the area of automated debugging. Nineteen automated debugging systems are analyzed. Thirteen existing automated debugging techniques are briefly evaluated from a pragmatic point of view. The three underlying strategies are identified, namely verification with respect to specification, checking with respect to language knowledge and filtering with respect to symptom. The verification strategy compares the actual program with some formal specification of the intended program. The checking strategy looks for suspect places which do not comply with some explicit knowledge of the programming language. The filtering strategy assumes correct parts of the code which cannot be responsible for the error symptom. Assertion evaluation and algorithmic debugging are the most promising verification techniques. Some intrinsic limitations of the checking strategy makes it only a complementary, though helpful, debugging support. The slicing technique should be included in any debugger.
    [bibtex-key = duc93b] [bibtex-entry]


  3. S. Le Huitouze, P. Louvet, and O. Ridoux. Les grammaires logiques et $\lambda$Prolog. In Journées Francophones sur la Programmation en Logique, Nîmes, France, pages 93-108, 1993. Teknea. [WWW] Keyword(s): LambdaProlog, grammaires logiques, portée, représentation du contexte..
    Abstract:
    La plupart des systèmes Prolog proposent un formalisme de grammaire logique appelé DCG (Definite Clause Grammar), dont l'utilité est reconnue. Nous présentons deux nouveaux formalismes de grammaire logique appelé DCG' et lambda-HHG (higher-order Hereditary Harrop Grammar)---grammaires héréditaires de Harrop d'ordre supérieur) destinés à être utilisés dans les systèmes LambdaProlog. Les relations entre DCG, DCG', et lambda-HHG, d'une part, et entre Prolog et LambdaProlog, d'autre part, peuvent être résumées de la manière suivante. (1) Prolog, DCG et la traduction de DCG en Prolog sont classiques. (2) Miller propose l'évolution de Prolog à LambdaProlog, et Pereira, Pareschi et Miller montrent l'intérêt d'utiliser LambdaProlog pour le traitement de la langue naturelle. (3) Nous proposons une variante fortement typée de DCG (appelée) afin de pouvoir la traduire en LambdaProlog dans le système LambdaProlog. C'est un premier pas vers un formalisme plus élaboré. (4) lambda-HHG est à DCG ce que LambdaProlog est à Prolog. Ce formalisme combine les avantages d'être grammatical et de cacher les opération d'un analyseur (comme DCG), et d'avoir des termes d'ordre supérieur comme attributs et de proposer une approche logique à la représentation des contextes (comme LambdaProlog).
    [bibtex-key = lehuitouze:grammaires:jfpl:93] [bibtex-entry]


  4. S. Le Huitouze, P. Louvet, and O. Ridoux. Logic Grammars and $\lambda$Prolog. In D.S. Warren, editor, 10th Int. Conf. Logic Programming, pages 64-79, 1993. MIT Press. [WWW] Keyword(s): LambdaProlog, logic grammars, scope, context handling in syntactic analysis..
    Abstract:
    A logic grammar formalism called DCG (Definite Clause Grammars), which has proved to be useful, is part of most Prolog implementations. We develop two new logic grammar formalisms called DCG' and lambda-HHG (higher-order Hereditary Harrop Grammars) that can be used in LambdaProlog implementations. The relations between DCG, DCG', and lambda-HHG, and Prolog and LambdaProlog can be summarized as follows: (1) The language Prolog, the DCG formalism, and the translation of DCG into Prolog by Prolog are classical. (2) The evolution from Prolog to LambdaProlog is due to Miller and the advantage of using LambdaProlog for doing natural language analysis is shown by Pereira, and Pareschi and Miller. (3) We propose a strongly typed variant of DCG (called DCG') for its translation into LambdaProlog by LambdaProlog. It is a first stage towards a more elaborate formalism. (4) A formalism that is to DCG what LambdaProlog is to Prolog is still missing, and also the way to translate it into LambdaProlog. Such a formalism combines the advantage of being grammatical and hiding the house-keeping operations (like DCG) and of having higher-order terms as attributes and providing a logical approach to context (like LambdaProlog). lambda-HHG is such a formalism.
    [bibtex-key = llr93] [bibtex-entry]


  5. P. Brisset and O. Ridoux. The Compilation of $\lambda$Prolog and its execution with MALI. Rapport de recherche 1831, INRIA, 1993. [WWW] [bibtex-key = brisset:compilation:inria:93] [bibtex-entry]


1992
  1. M. Ducassé, Y.-J. Lin, and L.Ü. Yalcinalp, editors. Proceedings of IJCSLP'92 Workshop on Logic Programming Environments, November 1992. Note: Technical Report TR 92-143, Case Western Reserve University, Cleveland. [bibtex-key = dly92] [bibtex-entry]


  2. M. Ducassé. An extendable trace analyser to support automated debugging. PhD thesis, University of Rennes I, France, June 1992. Note: European Doctorate.
    Abstract:
    The dissertation describes the innovative features of Opium, a high-level debugging environment for Prolog, designed and implemented at ECRC between 1985 and 1991. Debugging is a costly process, and automating it would significantly reduce the cost of software production and maintenance. However, it is unrealistic to aim at fully automating the task. In particular programmers have to understand rapidly changing situations, examining large amounts of data. In the current state of the art it is beyond the capabilities of computers to take the place of programmer's understanding. Nevertheless, computers can significantly help programmers to select the data to be analysed. The data used by program analysis in general is often restricted to the source code of the analysed programs. However, there is a complementary source of information, namely traces of program executions. An execution trace contains less general information than the program source, but it tells how the program behaves in a particular case. Furthermore, there are intrinsically dynamic aspects in a program which are best analysed at execution time, for example uses of read/write. These remarks suggested to build the automated debugging functionalities of Opium on top of an existing tracer, extending it to a general trace and source analyser. The most important features of Opium, not to be found in other debuggers, are as follows. - It provides a trace query language which is a solution to the ever growing command sets of other tracers. With two primitives plus Prolog, users can already specify more precise trace queries than with the hard coded commands of other tracers. - Opium is programmable and extendable. It is thus an environment where debugging strategies can be easily programmed and integrated. Some strategies are already implemented. - Abstract views of executions are proposed as a basis for automated debugging. They help users to understand the behaviours of programs by browsing through executions at a higher level than single steppers. Opium is fully implemented. More than 20 academic sites have recently requested Opium prototype, and some are actually implementing new abstract views.
    [bibtex-key = duc92] [bibtex-entry]


  3. Y. Bekkers, O. Ridoux, and L. Ungaro. Dynamic Memory Management for Sequential Logic Programming Languages. In Y. Bekkers and J. Cohen, editors, Int. Worshop on Memory Management, volume 637 of LNCS, pages 82-102, 1992. Springer-Verlag. [WWW] Keyword(s): Memory management, logic programming, garbage collection, usefulness logic..
    Abstract:
    Logic programming languages are becoming more complex with the introduction of new features such as constraints or terms with an equality theory. With this increase in complexity, they require more and more sophisticated memory management. This survey gives an insight into the memory management problems in sequential logic programming languages implementations; it also describes the presently know solutions. It is meant to be understood by non-specialists in logic programming with good knowledge of memory management in general. We first describe a "usefulness logic" for run-time objects. Usefulness logic defines non-garbage objects. Next, memory management systems are presented from the most trivial original run-time system, with no real concern for memory problems, to elaborated run-time systems with memory management closely observing the usefulness logic. Finally, the choice of a garbage collection technique is discussed in relation with logic programming specifities.
    [bibtex-key = bekkers:dynamic:iwmm:92] [bibtex-entry]


  4. P. Brisset and O. Ridoux. The Architecture of an Implementation of $\lambda$Prolog: Prolog/Mali. In Workshop on $\lambda$Prolog, Philadelphia, 1992. [WWW] Keyword(s): LambdaProlog, implementation, compilation, memory management..
    Abstract:
    LambdaProlog is a logic programming language accepting a more general clause form than standard Prolog (namely hereditary Harrop formulas instead of Horn formulas) and using simply typed lambda-terms as a term domain instead of first order terms. Despite these extensions, it is still amenable to goal-directed proofs and can still be given procedural semantics. However, the execution of LambdaProlog programs requires several departures from the standard resolution scheme. First, the augmented clause form causes the program (a set of clauses) and the signature (a set of constants) to be changeable, but in a very disciplined way. Second, the new term domain has a semi-decidable and infinitary unification theory, and it introduces the need for a beta-reduction operation at run-time. MALI is an abstract memory that is suitable for storing the search-state of depth-first search processes. Its main feature is its efficient memory management. We have used an original LambdaProlog-to-C translation: predicates are transformed into functions operating on several continuations. The compilation scheme is sometimes an adaptation of the standard Prolog scheme, but at other times it has to handle new features such as types, beta-reduction and delayed unification. Two keywords of this implementation are "sharing" and "folding" of representations. Sharing amounts to recognising that some representation already exists and reusing it. Folding amounts to recognising that two different representations represent the same thing and replacing one by the other. We assume a basic knowledge of Prolog and LambdaProlog.
    [bibtex-key = bo92b] [bibtex-entry]


  5. M. Ducassé. A general trace query mechanism based on Prolog. In M. Bruynooghe and M. Wirsing, editors, International Symposium on Programming Language Implementation and Logic Programming, volume 631 of Lecture Notes in Computer Science, pages 400-414, August 1992. Springer-Verlag. [WWW]
    Abstract:
    We present a general trace query language which is a solution to the ever growing command sets of other tracers. It provides all the required generality while being very simple and efficient. We model a program execution into a trace which is a stream of events. Execution events have a uniform representation, and can be analysed by Prolog programs. With this approach and thanks to the expressive power of Prolog, two high-level primitives plus Prolog are enough to provide a general trace query language. With a few optimizations this language can work on large executions without any loss of performance, if compared to traditional tracers. This paper describes the trace query mechanism from its high level specification down to some implementation details. The proposed model of trace query depends only on the sequentiality of the execution, and the principles behind the design of the optimizations do not depend on the traced language.
    [bibtex-key = duc92f] [bibtex-entry]


  6. M. Ducassé. A trace analyser to prototype explanations. In Proceedings of JICSLP'92 Workshop on Logic Programming Environments, Washington D.C., November 1992. Note: Technical Report TR 92-143, Case Western Reserve University, Cleveland.
    Abstract:
    Automated debugging and expert system explanations have in common to aim at helping people to understand executions. We have designed an extendable trace analyser for the purpose of automated debugging, which we propose to use to prototype expert system explanations. Two examples illustrate how simple it is to implement abstract tracing of executions and how easy it is to play with them.
    [bibtex-key = duc92b] [bibtex-entry]


  7. M. Ducassé. Analysis of failing Prolog executions. In Actes des Journées Francophones sur la Programmation Logique, Mai 1992.
    Abstract:
    The result of a Prolog execution can simply be ``no'', when the programmer is expecting something else. This symptom is typical of Prolog, and especially requires the help of an execution tracer to get clues of what the problem can be. We present a solution which helps programmers to understand how unexpected failures have occurred. We first propose a hierarchy of failing goals. We argue that there is one kind of leaf failures which is interesting to track at the first place. Then we give the algorithm for our leaf failure tracking and two examples illustrating its use.
    [bibtex-key = duc92c] [bibtex-entry]


  8. M. Ducassé. Opium: An advanced debugging system. In G. Comyn and N. Fuchs, editors, Proceedings of the Second Logic Programming Summer School, September 1992. Esprit Network of Excellence in Computational Logic COMPULOG-NET, Springer-Verlag, LNAI 636. [WWW]
    Abstract:
    The data used by program analysis in general is often restricted to the source code of the analysed programs. However, there is a complementary source of information, namely traces of program executions. Usual tracers, which extract this trace information, do not allow for general trace analysis. Opium, our debugger for Prolog, sets up a framework where program sources and traces of program executions can be jointly analysed. As the debugging process is heuristic and not all the debugging strategies have been identified so far, Opium is programmable. In particular, its trace query language gives more flexibility and more power than the hard coded command sets of usual tracers. This trace query language is based on Prolog. Opium is therefore both a helpful tool for Prolog and a nice application of Prolog. The most innovative extensions of Opium compute abstract views of Prolog executions to help users understand the behaviours of programs. In particular they help them understand how error symptoms have been produced. This article briefly recalls some general information about Opium. A debugging session is then commented in detail.
    [bibtex-key = duc92d] [bibtex-entry]


  9. P. Brisset and O. Ridoux. The Compilation of $\lambda$Prolog and its execution with MALI. Publication Interne 687, IRISA, 1992. [WWW] Keyword(s): LambdaProlog, implementation, compilation, memory management..
    Abstract:
    We present a compiled implementation of LambdaProlog that uses the abstract memory MALI for representing the execution state. LambdaProlog is a logic programming language allowing a more general clause form than Standard Prolog's (namely hereditary Harrop formulas instead of Horn formulas) and using simply typed lambda-terms as a term domain instead of first order terms. The augmented clause form causes the program (a set of clauses) and the signature (a set of constants) to be changeable in a very disciplined way. The new term domain has a semi-decidable and infinitary unification theory, and it introduces the need for a beta-reduction operation at run-time. MALI is an abstract memory that is suitable for storing the search-state of depth-first search processes. Its main feature is its efficient memory management. We have used an original LambdaProlog-to-C translation along which predicates are transformed into functions operating on continuations for handling failure and success in unifications, and changes in signatures and programs. Two keywords of this implementation are ``sharing'' and ``folding'' of representations. Sharing amounts to recognising that some representation already exists and to reuse it. Folding amounts to recognising that two different representations represent the same thing and to replace one by the other.
    [bibtex-key = bo92a] [bibtex-entry]


1991
  1. M. Ducassé and G. Ferrand, editors. Proceedings of ICLP'91 Workshop on Logic Programming Environments, June 1991. Note: Technical Report, University of Orléans, France, LIFO N 91-61. [bibtex-key = df91] [bibtex-entry]


  2. M. Ducassé and A.-M. Emde. Opium: a debugging environment for Prolog development and debugging research. ACM Software Engineering Notes, 16(1):54-59, January 1991. Note: Demonstration presented at the Fourth Symposium on Software Development Environments. Keyword(s): demo.
    Abstract:
    Opium is an extensible debugging environment for PROLOG providing high-level debugging facilities for programmers and debugging experts. In the design of debuggers there are two tasks which are often mixed, extraction and analysis of debugging information. The aim of the extraction task is to collect the whole debugging information so that users do not miss any important information about their program. On the other hand, the aim of the analysis task is to restrict in an accurate way the amount of debugging information shown to the user so that the latter has to examine only the relevant parts. This task clearly depends on the debugging situation and, to our point of view, there is no general restriction which can be done a priori. However, the two tasks are usually mixed and hard-coded, the result is that not enough relevant information and too much useless information is displayed. In Opium the two tasks are clearly separated. The extraction module collects the whole debugging information (execution trace and program source) which is then available for the analysis module. The presentation concentrates on the analysis module, discussing the main aspects of Opium: programmability, high-level debugging, extensibility mechanisms, meta-debugging, support for end-users and debugging experts.
    [bibtex-key = de91d] [bibtex-entry]


  3. P. Brisset and O. Ridoux. Naïve Reverse Can Be Linear. In K. Furukawa, editor, 8th Int. Conf. Logic Programming, pages 857-870, 1991. MIT Press. [WWW] Keyword(s): LambdaProlog, implementation, function-lists, higher-order unification..
    Abstract:
    We propose a new implementation of logic programming with higher-order terms. In order to illustrate the properties of our implementation, we apply the coding of lists as functions to the context of logic programming. As a side-effect, we show that higher-order unification is a good tool for manipulating the function-lists. It appears that the efficiency of the program thus obtained relies critically upon the implementation of higher-order operations (unification and reduction). In particular, we show that a good choice for data-structures and reduction strategy yields a linear naïve reverse.
    [bibtex-key = brisset:naivereverse:iclp:91] [bibtex-entry]


  4. M. Ducassé. Abstract views of Prolog executions in Opium. In V. Saraswat and K. Ueda, editors, Proceedings of the International Logic Programming Symposium, San Diego, USA, pages 18-32, October 1991. MIT Press. [WWW]
    Abstract:
    Opium is a system for analysing and debugging Prolog programs. Its kernel comprises an execution tracer and a programming language with a full set of primitives for trace and source analysis. In this paper we show the power of Opium for supporting abstract views of Prolog executions. Abstract views give high-level points of view about the execution. They filter out irrelevant details; they restructure the remaining information; and they compact it so that the amount of information given at each step has a reasonable size. The examples of abstract views given in the following are a goal execution profile, some data abstractions, an instantiation profile, a failure analysis and a kind of explanation for an expert system written in Prolog.
    [bibtex-key = duc91c] [bibtex-entry]


  5. M. Ducassé and A.-M. Emde. A High-level Debugging Environment for Prolog. Opium User's Manual. Technical Report TR-LP-60, ECRC, May 1991. [WWW] [bibtex-key = de91] [bibtex-entry]


1988
  1. M. Ducassé. Opium+, a Meta-debugger for Prolog. In Proceedings of the European Conference on Artificial Intelligence, Munich, pages 272-277, August 1988. [bibtex-key = ducasse88f] [bibtex-entry]


  2. M. Ducassé and A.-M. Emde. A Review of Automated Debugging Systems: Knowledge, Strategies, and Techniques. In Proceedings of the 10th International Conference on Software Engineering, pages 162-171, 1988. IEEE Press. [bibtex-key = ducasse88] [bibtex-entry]



BACK TO INDEX




Disclaimer:

This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All person copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.

Les documents contenus dans ces répertoires sont rendus disponibles par les auteurs qui y ont contribué en vue d'assurer la diffusion à temps de travaux savants et techniques sur une base non-commerciale. Les droits de copie et autres droits sont gardés par les auteurs et par les détenteurs du copyright, en dépit du fait qu'ils présentent ici leurs travaux sous forme électronique. Les personnes copiant ces informations doivent adhérer aux termes et contraintes couverts par le copyright de chaque auteur. Ces travaux ne peuvent pas être rendus disponibles ailleurs sans la permission explicite du détenteur du copyright.




Last modified: Fri Jan 28 13:48:42 2022
Author: ferre.


This document was translated from BibTEX by bibtex2html