Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Edgard Marx is active.

Publication


Featured researches published by Edgard Marx.


international conference on semantic systems | 2014

Towards an open question answering architecture

Edgard Marx; Axel-Cyrille Ngonga Ngomo; Konrad Höffner; Jens Lehmann; Sören Auer

Billions of facts pertaining to a multitude of domains are now available on the Web as RDF data. However, accessing this data is still a difficult endeavour for non-expert users. In order to meliorate the access to this data, approaches imposing minimal hurdles to their users are required. Although many question answering systems over Linked Data have being proposed, retrieving the desired data is still significantly challenging. In addition, developing and evaluating question answering systems remains a very complex task. To overcome these obstacles, we present a modular and extensible open-source question answering framework. We demonstrate how the framework can be used by integrating two state-of-the-art question answering systems. As a result our evaluation shows that overall better results can be achieved by the use of combination rather than individual stand-alone versions.


Sprachwissenschaft | 2017

Survey on challenges of Question Answering in the Semantic Web

Konrad Höffner; Sebastian Walter; Edgard Marx; Jens Lehmann; Axel-Cyrille Ngonga Ngomo

Semantic Question Answering (SQA) removes two major access requirements to the Semantic Web: the mastery of a formal query language like SPARQL and knowledge of a specific vocabulary. Because of the complexity of natural language, SQA presents difficult challenges and many research opportunities. Instead of a shared effort, however, many essential components are redeveloped, which is an inefficient use of researcher’s time and resources. This survey analyzes 62 different SQA systems, which are systematically and manually selected using predefined inclusion and exclusion criteria, leading to 72 selected publications out of 1960 candidates. We identify common challenges, structure solutions, and provide recommendations for future systems. This work is based on publications from the end of 2010 to July 2015 and is also compared to older but similar surveys.


international world wide web conferences | 2015

ROCKER: A Refinement Operator for Key Discovery

Tommaso Soru; Edgard Marx; Axel-Cyrille Ngonga Ngomo

The Linked Data principles provide a decentral approach for publishing structured data in the RDF format on the Web. In contrast to structured data published in relational databases where a key is often provided explicitly, finding a set of properties that allows identifying a resource uniquely is a non-trivial task. Still, finding keys is of central importance for manifold applications such as resource deduplication, link discovery, logical data compression and data integration. In this paper, we address this research gap by specifying a refinement operator, dubbed ROCKER, which we prove to be finite, proper and non-redundant. We combine the theoretical characteristics of this operator with two monotonicities of keys to obtain a time-efficient approach for detecting keys, i.e., sets of properties that describe resources uniquely. We then utilize a hash index to compute the discriminability score efficiently. Therewith, we ensure that our approach can scale to very large knowledge bases. Results show that ROCKER yields more accurate results, has a comparable runtime, and consumes less memory w.r.t. existing state-of-the-art techniques.


ieee international conference semantic computing | 2013

Large-Scale RDF Dataset Slicing

Edgard Marx; Saeedeh Shekarpour; Sören Auer; Axel-Cyrille Ngonga Ngomo

In the last years an increasing number of structured data was published on the Web as Linked Open Data (LOD). Despite recent advances, consuming and using Linked Open Data within an organization is still a substantial challenge. Many of the LOD datasets are quite large and despite progress in RDF data management their loading and querying within a triple store is extremely time-consuming and resource-demanding. To overcome this consumption obstacle, we propose a process inspired by the classical Extract-Transform-Load (ETL) paradigm. In this article, we focus particularly on the selection and extraction steps of this process. We devise a fragment of SPARQL dubbed SliceSPARQL, which enables the selection of well-defined slices of datasets fulfilling typical information needs. SliceSPARQL supports graph patterns for which each connected sub graph pattern involves a maximum of one variable or IRI in its join conditions. This restriction guarantees the efficient processing of the query against a sequential dataset dump stream. As a result our evaluation shows that dataset slices can be generated an order of magnitude faster than by using the conventional approach of loading the whole dataset into a triple store and retrieving the slice by executing the query against the triple stores SPARQL endpoint.


knowledge acquisition, modeling and management | 2016

ACRyLIQ: Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment

Umair ul Hassan; Amrapali Zaveri; Edgard Marx; Edward Curry; Jens Lehmann

Crowdsourcing has emerged as a powerful paradigm for quality assessment and improvement of Linked Data. A major challenge of employing crowdsourcing, for quality assessment in Linked Data, is the cold-start problem: how to estimate the reliability of crowd workers and assign the most reliable workers to tasks? We address this challenge by proposing a novel approach for generating test questions from DBpedia based on the topics associated with quality assessment tasks. These test questions are used to estimate the reliability of the new workers. Subsequently, the tasks are dynamically assigned to reliable workers to help improve the accuracy of collected responses. Our proposed approach, ACRyLIQ, is evaluated using workers hired from Amazon Mechanical Turk, on two real-world Linked Data datasets. We validate the proposed approach in terms of accuracy and compare it against the baseline approach of reliability estimate using gold-standard task. The results demonstrate that our proposed approach achieves high accuracy without using gold-standard task.


ieee international conference semantic computing | 2017

Torpedo: Improving the State-of-the-Art RDF Dataset Slicing

Edgard Marx; Saeedeh Shekarpour; Tommaso Soru; Adrian M.P. Braşoveanu; Muhammad Saleem; Ciro Baron; Albert Weichselbraun; Jens Lehmann; Axel-Cyrille Ngonga Ngomo; Sören Auer

Over the last years, the amount of data published as Linked Data on the Web has grown enormously. In spite of the high availability of Linked Data, organizations still encounter an accessibility challenge while consuming it. This is mostly due to the large size of some of the datasets published as Linked Data. The core observation behind this work is that a subset of these datasets suffices to address the needs of most organizations. In this paper, we introduce Torpedo, an approach for efficiently selecting and extracting relevant subsets from RDF datasets. In particular, Torpedo adds optimization techniques to reduce seek operations costs as well as the support of multi-join graph patterns and SPARQL FILTERs that enable to perform a more granular data selection. We compare the performance of our approach with existing solutions on nine different queries against four datasets. Our results show that our approach is highly scalable and is up to 26% faster than the current state-of-the-art RDF dataset slicing approach.


international conference on semantic systems | 2016

DBtrends: Exploring Query Logs for Ranking RDF Data

Edgard Marx; Amrapali Zaveri; Diego Moussallem; Sandro Rautenberg

Many ranking methods have been proposed for RDF data. These methods often use the structure behind the data to measure its importance. Recently, some of these methods have started to explore information from other sources such as the Wikipedia page graph for better ranking RDF data. In this work, we propose DBtrends, a ranking function based on query logs. We extensively evaluate the application of different ranking functions for entities, classes, and properties across two different countries as well as their combination. Thereafter, we propose MIXED-RANK, a ranking function that combines DBtrends with the best-evaluated entity ranking function. We show that: (i) MIXED-RANK outperforms state-of-the-art entity ranking functions, and; (ii) query logs can be used to improve RDF ranking functions.


international conference on semantic systems | 2015

LODFlow: a workflow management system for linked data processing

Sandro Rautenberg; Ivan Ermilov; Edgard Marx; Sören Auer; Axel-Cyrille Ngonga Ngomo

The extraction and maintenance of Linked Data datasets is a cumbersome, time-consuming and resource-intensive activity. The cost for producing Linked Data can be reduced by a workflow management system, which describes plans to systematically support the lifecycle of RDF datasets. We present the LODFlow Linked Data Workflow Management System, which provides an environment for planning, executing, reusing, and documenting Linked Data workflows. The LODFlow approach is based on a comprehensive knowledge model for describing the workflows and a workflow execution engine supporting systematic workflow execution, reporting, and exception handling. The environment was evaluated in a large-scale real-world use case. As result, LODFlow supports Linked Data engineers to systematically plan, execute and assess Linked Data production and maintenance workflows, thus improving efficiency, ease-of-use, reproducibility, reuseability and provenance. The environment was evaluated in a large-scale real-world use case. As result, LODFlow supports Linked Data engineers to systematically plan, execute and assess Linked Data production and maintenance workflows, thus improving efficiency, ease-of-use, reproducibility, reuseability and provenance.


International Journal of Semantic Computing | 2013

TOWARDS AN EFFICIENT RDF DATASET SLICING

Edgard Marx; Tommaso Soru; Saeedeh Shekarpour; Sören Auer; Axel-Cyrille Ngonga Ngomo; Karin Koogan Breitman

Over the last years, a considerable amount of structured data has been published on the Web as Linked Open Data (LOD). Despite recent advances, consuming and using Linked Open Data within an organization is still a substantial challenge. Many of the LOD datasets are quite large and despite progress in Resource Description Framework (RDF) data management their loading and querying within a triple store is extremely time-consuming and resource-demanding. To overcome this consumption obstacle, we propose a process inspired by the classical Extract-Transform-Load (ETL) paradigm. In this article, we focus particularly on the selection and extraction steps of this process. We devise a fragment of SPARQL Protocol and RDF Query Language (SPARQL) dubbed SliceSPARQL, which enables the selection of well-defined slices of datasets fulfilling typical information needs. SliceSPARQL supports graph patterns for which each connected subgraph pattern involves a maximum of one variable or Internationalized resource identifier (IRI) in its join conditions. This restriction guarantees the efficient processing of the query against a sequential dataset dump stream. Furthermore, we evaluate our slicing approach on three different optimization strategies. Results show that dataset slices can be generated an order of magnitude faster than by using the conventional approach of loading the whole dataset into a triple store.


ieee international conference semantic computing | 2017

KBox — Transparently Shifting Query Execution on Knowledge Graphs to the Edge

Edgard Marx; Ciro Baron; Tommaso Soru; Sören Auer

The Semantic Web architecture choices lead to a tremendous amount of information published on the Web. However, to query and build applications on top of Linked Open Data is still significantly challenging. In this work, we present Knowledge Box (KBox), an approach for transparently shifting query execution on Knowledge Graphs to the edge. We show that our approach makes the consumption of Knowledge Graphs more reliable and faster than the formerly introduced methods. In practice, KBox demands less time and resources to setup than traditional approaches, while being twice as fast as SPARQL endpoints, even when serving a single client.

Collaboration


Dive into the Edgard Marx's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sandro Rautenberg

Midwestern State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge