Pedro J. Abad
University of Huelva
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Pedro J. Abad.
IEEE Transactions on Knowledge and Data Engineering | 2016
Iñaki Fernández de Viana; Pedro J. Abad; José Luis Álvarez; José Luis Arjona
Wrappers are pieces of software used to extract data from websites and structure them for further application processing. Unfortunately, websites are continuously evolving and structural changes happen with no forewarning, which usually results in wrappers working incorrectly. Thus, wrappers maintenance is necessary for detecting whether wrapper is extracting erroneous data. The solution consists of using verification models to detect whether wrapper output is statistically similar to the output produced by the wrapper itself when it was successfully invoked in the past. Current proposals present some weaknesses, as the data used to build these models are supposed to be homogeneous, independent, or representative enough, or following a single predefined mathematical model. In this paper, we present MAVE, a novel multilevel wrapper verification system that is based on one-class classification techniques to overcome previous weaknesses. The experimental results show that our proposal outperforms accuracy of current solutions.
Expert Systems With Applications | 2016
I. Fernández de Viana; Pedro J. Abad; José Luis Álvarez; José Luis Arjona
First application of Ant Colony Metaheuristic to verify information extracted by web wrappers.New multilevel verification system that improves the results achieved by current techniques.Enumeration of current techniques weakness.Reformulation of wrapper verification problem as a combinational optimization problem.Applying non-parametric testing techniques to ascertain the statistical significance among results. Wrappers are pieces of software used to extract data from websites and structure them for further application processing. Unfortunately, websites are continuously evolving and structural changes happen with no forewarning, which usually results in wrappers working incorrectly. Thus, wrappers maintenance is necessary for detecting whether wrapper is extracting erroneous data. The solution consists of using verification models to detect whether wrapper output is statistically similar to the output produced by the wrapper itself when it was successfully invoked in the past. Current proposals present some weaknesses, as the data used to build these models are supposed to be homogeneous or that the features of this data set can be mapped to an n-dimensional space of independent dimensions when there is a correlation among their features. In this paper, a new verification system based on the Best-Worst Ant System (BWAS) is presented to overcome previous weaknesses. The experimental results show an accuracy improvement of 7.5% over current solutions.
ISD | 2014
Diana Borrego; Irene Barba; Pedro J. Abad
Since the accurate management of business processes is receiving increasing attention, conformance checking, i.e., verifying whether the observed behaviour matches a modelled behaviour, is becoming more and more critical. Moreover, declarative languages are more frequently used to provide an increased flexibility. However, little work has been conducted to deal with conformance checking of declarative models. Furthermore, only control-flow perspective is usually considered although other perspectives (e.g., data) are crucial. In addition, most approaches exclusively check the conformance without providing any diagnostics. To enhance the accurate management of flexible business processes, this work presents a constraint-based approach for conformance checking over declarative models (including control-flow and data perspectives) and for providing related diagnosis.
practical applications of agents and multi agent systems | 2011
F. J. Fernández; José Luis Álvarez; Pedro J. Abad; Patricia Jiménez
The Web is the largest repository of useful information available for human users, but it is usual that Web Pages do not provide an API to get access to its information automatically. In order to solve this problem, Information Extractors are developed. We present a new methodology to induce Information Extractors from the Web. It is based on rendering HTML elements in the Web browser. The methodology uses a KDD process to mining a dataset with features of the elements in the Web page. An experimentation over 10 web sites has been made and the results show the effectiveness of the methodology.
IEEE Latin America Transactions | 2010
I. F. de Viana; José Luis Arjona; José Luis Álvarez; Pedro J. Abad
Reduce maintenance costs of Enterprise Application Integration (EAI) solutions becomes a challenge when we are trying to integrate friendly web applications. This problem can be solved using automated systems that allow to navigate, extract, structure and verify relevant information. The extracted information is characterized using complex models that later are used to check whether the information is valid. In this paper we will demonstrate empirically that feature selection techniques simplify and improve the models obtained for verification of information.
IEEE Latin America Transactions | 2010
I. F. de Viana; José Luis Arjona; José Luis Álvarez; Pedro J. Abad
Reduce maintenance costs of Enterprise Application Integration (EAI) solutions becomes a challenge when we are trying to integrate friendly web applications. This problem can be solved using automated systems that allow to navigate, extract, structure and verify relevant information. The extracted information is characterized using complex models that later are used to check whether the information is valid. In this paper we will demonstrate empirically that feature selection techniques simplify and improve the models obtained for verification of information.
computational intelligence for modelling, control and automation | 2005
Pedro J. Abad; Antonio J. Suárez; Rafael M. Gasca; Juan Antonio Ortega
This paper proposes a methodology to diagnose a transient state of a dynamic system using boosting. The methodology is composed by two steps: one off-line process and another on-line process. The off-line phase begins gathering data from the system, both when it is running free of fault and when the system is running in each fault mode. A segmentation and normalization algorithm is used to reduce the large amount of gathered data. The final step is the generation of a decision tree by a classification tool. The boosting technique is used with the aim of improving the classification results. The on-line process of the methodology consists of evaluating a new reading of the system sensors with the generated decision trees. The diagnosis of the system is the result of this evaluation which has very low computational cost due to the simplicity of the decision trees. Also, the implementation cost is very low due to this simplicity
Archive | 2002
Pedro J. Abad; Antonio J. Suárez; Rafael M. Gasca; Juan Antonio Ortega
international conference on software and data technologies | 2011
Iñaki Fernández de Viana; Pedro J. Abad; José Luis Álvarez; José Luis Arjona
iberian conference on information systems and technologies | 2011
Iñaki Fernández de Viana; Pedro J. Abad; José Luis Álvarez; José Luis Arjona