Manuel Atencia
University of Grenoble
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Manuel Atencia.
knowledge acquisition, modeling and management | 2012
Manuel Atencia; Jérôme David; François Scharffe
This paper introduces a method for analyzing web datasets based on key dependencies. The classical notion of a key in relational databases is adapted to RDF datasets. In order to better deal with web data of variable quality, the definition of a pseudo-key is presented. An RDF vocabulary for representing keys is also provided. An algorithm to discover keys and pseudo-keys is described. Experimental results show that even for a big dataset such as DBpedia, the runtime of the algorithm is still reasonable. Two applications are further discussed: (i) detection of errors in RDF datasets, and (ii) datasets interlinking.
international semantic web conference | 2012
Manuel Atencia; Alexander Borgida; Jérôme Euzenat; Chiara Ghidini; Luciano Serafini
Ontology mappings are often assigned a weight or confidence factor by matchers. Nonetheless, few semantic accounts have been given so far for such weights. This paper presents a formal semantics for weighted mappings between different ontologies. It is based on a classificational interpretation of mappings: if O1 and O2 are two ontologies used to classify a common set X, then mappings between O1 and O2 are interpreted to encode how elements of X classified in the concepts of O1 are re-classified in the concepts of O2, and weights are interpreted to measure how precise and complete re-classifications are. This semantics is justifiable by extensional practice of ontology matching. It is a conservative extension of a semantics of crisp mappings. The paper also includes properties that relate mapping entailment with description logic constructors.
Journal of Web Semantics | 2012
Manuel Atencia; W. Marco Schorlemmer
We tackle the problem of semantic heterogeneity in the context of agent communication and argue that solutions based solely on ontologies and ontology matching do not capture adequately the richness of semantics as it arises in dynamic and open multiagent systems. Current solutions to the semantic heterogeneity problem in distributed systems usually do not address the contextual nuances of the interaction underlying an agent communication. The meaning an agent attaches to its utterances is, in our view, very relative to the particular dialogue in which it may be engaged, and that the interaction model specifying its dialogical structure and its unfolding should not be left out of the semantic alignment mechanism. In this article we provide the formal foundation of a novel, interaction-based approach to semantic alignment, drawing from a mathematical construct inspired from category theory that we call the communication product. In addition, we describe a simple alignment protocol which, combined with a probabilistic matching mechanism, endows an agent with the capacity of bootstrapping - by repeated successful interaction - the basic semantic relationship between its local vocabulary and that of another agent. We have also implemented the alignment technique based on this approach and prove its viability by means of an abstract experimentation and a thorough statistical analysis.
european conference on artificial intelligence | 2014
Manuel Atencia; Jérôme David; Jérôme Euzenat
Links are important for the publication of RDF data on the web. Yet, establishing links between data sets is not an easy task. We develop an approach for that purpose which extracts weak linkkeys. Linkkeys extend the notion of a key to the case of different data sets. They are made of a set of pairs of properties belonging to two different classes. A weak linkkey holds between two classes if any resources having common values for all of these properties are the same resources. An algorithm is proposed to generate a small set of candidate linkkeys. Depending on whether some of the, valid or invalid, links are known, we define supervised and non supervised measures for selecting the appropriate linkkeys. The supervised measures approximate precision and recall, while the non supervised measures are the ratio of pairs of entities a linkkey covers (coverage), and the ratio of entities from the same data set it identifies (discrimination). We have experimented these techniques on two data sets, showing the accuracy and robustness of both approaches.
international conference on conceptual structures | 2014
Manuel Atencia; Michel Chein; Madalina Croitoru; Jérôme David; Michel Leclère; Nathalie Pernelle; Fatiha Saïs; François Scharffe; Danai Symeonidou
Many techniques were recently proposed to automate the linkage of RDF datasets. Predicate selection is the step of the linkage process that consists in selecting the smallest set of relevant predicates needed to enable instance comparison. We call keys this set of predicates that is analogous to the notion of keys in relational databases. We explain formally the different assumptions behind two existing key semantics. We then evaluate experimentally the keys by studying how discovered keys could help dataset interlinking or cleaning. We discuss the experimental results and show that the two different semantics lead to comparable results on the studied datasets.
european conference on artificial intelligence | 2016
Mustafa Al-Bakri; Manuel Atencia; Jérôme David; Steffen Lalande; Marie-Christine Rousset
Discovering whether or not two URIs described in Linked Data -- in the same or different RDF datasets -- refer to the same real-world entity is crucial for building applications that exploit the cross-referencing of open data. A major challenge in data interlinking is to design tools that effectively deal with incomplete and noisy data, and exploit uncertain knowledge. In this paper, we model data interlinking as a reasoning problem with uncertainty. We introduce a probabilistic framework for modelling and reasoning over uncertain RDF facts and rules that is based on the semantics of probabilistic Datalog. We have designed an algorithm, ProbFR, based on this framework. Experiments on real-world datasets have shown the usefulness and effectiveness of our approach for data linkage and disambiguation.
knowledge acquisition, modeling and management | 2012
Mustafa Al-Bakri; Manuel Atencia; Marie-Christine Rousset
Virtual online communities (social networks, wikis. . . ) are becoming the major usage of the web. The freedom they give to publish and access information is attracting many web users. However, this freedom is filling up the web with varied information and viewpoints. This raises important issues that concern privacy and trust. Due to their decentralised nature peer-to-peer (P2P) systems provide a partial solution for the privacy problem: each user (peer) can keep control on her own data by storing it locally and by deciding the access they want to give to other peers. We focus on semantic P2P systems in which peers annotate their resources (documents, videos, photos, services) using ontologies.
international semantic web conference | 2011
Manuel Atencia; Jérôme Euzenat; Giuseppe Pirrò; Marie-Christine Rousset
11th ISWC workshop on ontology matching (OM) | 2016
Maroua Gmati; Manuel Atencia; Jérôme Euzenat
Archive | 2015
Jérôme David; Jérôme Euzenat; Manuel Atencia