Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Álvaro Rodrigo is active.

Publication


Featured researches published by Álvaro Rodrigo.


Advances in Multilingual and Multimodal Information Retrieval | 2008

Overview of the Answer Validation Exercise 2007

Anselmo Peñas; Álvaro Rodrigo; Felisa Verdejo

The Answer Validation Exercise at the Cross Language Evaluation Forum is aimed at developing systems able to decide whether the answer of a Question Answering system is correct or not. We present here the exercise description, the changes in the evaluation methodology with respect to the first edition, and the results of this second edition (AVE 2007). The changes in the evaluation methodology had two objectives: the first one was to quantify the gain in performance when more sophisticated validation modules are introduced in QA systems. The second objective was to bring systems based on Textual Entailment to the Automatic Hypothesis Generation problem which is not part itself of the Recognising Textual Entailment (RTE) task but a need of the Answer Validation setting. 9 groups have participated with 16 runs in 4 different languages. Compared with the QA systems, the results show an evidence of the potential gain that more sophisticated AV modules introduce in the task of QA.


cross language evaluation forum | 2009

Overview of ResPubliQA 2009: question answering evaluation over European legislation

Anselmo Peñas; Pamela Forner; Richard F. E. Sutcliffe; Álvaro Rodrigo; Corina Forăscu; Iñaki Alegria; Danilo Giampiccolo; Nicolas Moreau; Petya Osenova

This paper describes the first round of ResPubliQA, a Question Answering (QA) evaluation task over European legislation, proposed at the Cross Language Evaluation Forum (CLEF) 2009. The exercise consists of extracting a relevant paragraph of text that satisfies completely the information need expressed by a natural language question. The general goals of this exercise are (i) to study if the current QA technologies tuned for newswire collections and Wikipedia can be adapted to a new domain (law in this case); (ii) to move to a more realistic scenario, considering people close to law as users, and paragraphs as system output; (iii) to compare current QA technologies with pure Information Retrieval (IR) approaches; and (iv) to introduce in QA systems the Answer Validation technologies developed in the past three years. The paper describes the task in more detail, presenting the different types of questions, the methodology for the creation of the test sets and the new evaluation measure, and analyzing the results obtained by systems and the more successful approaches. Eleven groups participated with 28 runs. In addition, we evaluated 16 baseline runs (2 per language) based only in pure IR approach, for comparison purposes. Considering accuracy, scores were generally higher than in previous QA campaigns.


Journal of Logic and Computation | 2008

Testing the Reasoning for Question Answering Validation

Anselmo Peñas; Álvaro Rodrigo; Valentín Sama; Felisa Verdejo

Question answering (QA) is a task that deserves more collaboration between natural language processing (NLP) and knowledge representation (KR) communities, not only to introduce reasoning when looking for answers or making use of answer type taxonomies and encyclopaedic knowledge, but also, as discussed here, for answer validation (AV), that is to say, to decide whether the responses of a QA system are correct or not. This was one of the motivations for the first Answer Validation Exercise at CLEF 2006 (AVE 2006). The starting point for the AVE 2006 was the reformulation of the answer validation as a recognizing textual entailment (RTE) problem, under the assumption that a hypothesis can be automatically generated instantiating a hypothesis pattern with a QA system answer. The test collections that we developed in seven different languages at AVE 2006 are specially oriented to the development and evaluation of answer validation systems. We show in this article the methodology followed for developing these collections taking advantage of the human assessments already made in the evaluation of QA systems. We also propose an evaluation framework for AV linked to a QA evaluation track. We quantify and discuss the source of errors introduced by the reformulation of the answer validation problem in terms of textual entailment (around 2%, in the range of inter-annotator disagreement). We also show the evaluation results of the first answer validation exercise at CLEF 2006 where 11 groups have participated with 38 runs in seven different languages. The most extensively used techniques were Machine Learning and overlapping measures, but systems with broader knowledge resources and richer representation formalisms obtained the best results.


cross language evaluation forum | 2006

The effect of entity recognition on answer validation

Álvaro Rodrigo; Anselmo Peñas; Jesús Herrera; Felisa Verdejo

The Answer Validation Exercise (AVE) 2006 is aimed at evaluating systems able to decide whether the responses of a Question Answering (QA) system are correct or not. Since most of the questions and answers contain entities, the use of a textual entailment relation between entities is studied here for the task of Answer Validation. We present some experiments concluding that the entity entailment relation is a feature that improves a SVM based classifier close to the best result in AVE 2006.


cross language evaluation forum | 2013

QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation

Anselmo Peñas; Eduard H. Hovy; Pamela Forner; Álvaro Rodrigo; Richard F. E. Sutcliffe; Roser Morante

This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011---2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjunction with background text collections provided by the organization. Four different tasks have been organized during these years: Main Task, Processing Modality and Negation for Machine Reading, Machine Reading of Biomedical Texts about Alzheimers disease, and Entrance Exams. This paper describes their motivation, their goals, their methodology for preparing the data sets, their background collections, their metrics used for the evaluation, and the lessons learned along these three years.


cross language evaluation forum | 2008

UNED at Answer Validation Exercise 2007

Álvaro Rodrigo; Anselmo Peñas; Felisa Verdejo

The objective of the Answer Validation Exercise (AVE) 2007 is to develop systems able to decide if the answer to a question is correct or not. Since it is expected that a high percentage of the answers, questions and supporting snippets contain named entities, the paper presents a method for validating answers that uses only information about named entities. The promising results encourage us to improve the system and use it as a component of other systems.


cross language evaluation forum | 2009

Information retrieval baselines for the ResPubliQA task

Joaquín Pérez-Iglesias; Guillermo Garrido; Álvaro Rodrigo; Lourdes Araujo; Anselmo Peñas

The baselines proposed for the ResPubliQA 2009 task are described in this paper. The main aim for designing these baselines was to test the performance of a pure Information Retrieval approach on this task. Two baselines were run for each of the eight languages of the task. Both baselines used the Okapi-BM25 ranking function, with and without a stemming. In this paper we extend the previous baselines comparing the BM25 model with Vector Space Model performance on this task. The results prove that BM25 outperforms VSM for all cases.


language resources and evaluation | 2012

Evaluating question answering validation as a classification problem

Álvaro Rodrigo; Anselmo Peñas; Felisa Verdejo

Formulating Question Answering Validation as a classification problem facilitates the introduction of Machine Learning techniques to improve the overall performance of Question Answering systems. The different proportion of positive and negative examples in the evaluation collections has led to the use of measures based on precision and recall. However, an evaluation based on the analysis of Receiver Operating Characteristic (ROC) space is sometimes preferred in classification with unbalanced collections. In this article we compare both evaluation approaches according to their rationale, their stability, their discrimination power and their adequacy to the particularities of the Answer Validation task.


cross-language evaluation forum | 2009

Approaching question answering by means of paragraph validation

Álvaro Rodrigo; Joaquín Pérez-Iglesias; Anselmo Peñas; Guillermo Garrido; Lourdes Araujo

In this paper we describe the QA system developed for taking part in Res-PubliQA 2009. Our system was composed by an IR phase focused on improving QA results, a validation step for removing paragraphs that are not promising and a module based on ngrams overlapping for selecting the final answer. Furthermore, a selection module that uses lexical entailment in combination with ngrams overlapping was developed in English. The IR module achieved very promising results that were improved by the ngram ranking. Moreover, the ranking was slightly improved when lexical entailment was used.


meeting of the association for computational linguistics | 2007

Experiments of UNED at the Third Recognising Textual Entailment Challenge

Álvaro Rodrigo; Anselmo Peñas; Jesús Herrera; Felisa Verdejo

This paper describes the experiments developed and the results obtained in the participation of UNED in the Third Recognising Textual Entailment (RTE) Challenge. The experiments are focused on the study of the effect of named entities in the recognition of textual entailment. While Named Entity Recognition (NER) provides remarkable results (accuracy over 70%) for RTE on QA task, IE task requires more sophisticated treatment of named entities such as the identification of relations between them.

Collaboration


Dive into the Álvaro Rodrigo's collaboration.

Top Co-Authors

Avatar

Anselmo Peñas

National University of Distance Education

View shared research outputs
Top Co-Authors

Avatar

Pamela Forner

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Guillermo Garrido

National University of Distance Education

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lourdes Araujo

National University of Distance Education

View shared research outputs
Top Co-Authors

Avatar

Joaquín Pérez-Iglesias

National University of Distance Education

View shared research outputs
Top Co-Authors

Avatar

Corina Forascu

Alexandru Ioan Cuza University

View shared research outputs
Top Co-Authors

Avatar

Petya Osenova

Bulgarian Academy of Sciences

View shared research outputs
Researchain Logo
Decentralizing Knowledge