José Luis Vázquez-Poletti
Complutense University of Madrid
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by José Luis Vázquez-Poletti.
grid computing | 2012
Alberto Núñez; José Luis Vázquez-Poletti; Agustín C. Caminero; Gabriel G. Castañé; Jesús Carretero; Ignacio Martín Llorente
Simulation techniques have become a powerful tool for deciding the best starting conditions on pay-as-you-go scenarios. This is the case of public cloud infrastructures, where a given number and type of virtual machines (in short VMs) are instantiated during a specified time, being this reflected in the final budget. With this in mind, this paper introduces and validates iCanCloud, a novel simulator of cloud infrastructures with remarkable features such as flexibility, scalability, performance and usability. Furthermore, the iCanCloud simulator has been built on the following design principles: (1) it’s targeted to conduct large experiments, as opposed to others simulators from literature; (2) it provides a flexible and fully customizable global hypervisor for integrating any cloud brokering policy; (3) it reproduces the instance types provided by a given cloud infrastructure; and finally, (4) it contains a user-friendly GUI for configuring and launching simulations, that goes from a single VM to large cloud computing systems composed of thousands of machines.
Future Generation Computer Systems | 2013
Rizwan Mian; Patrick Martin; José Luis Vázquez-Poletti
Data analytics applications are well-suited for a cloud environment. In this paper we examine the problem of provisioning resources in a public cloud to execute data analytic workloads. The goal of our provisioning method is to determine the most cost-effective configuration for a given data analytic workload. Provisioning a workload in a public cloud environment faces several challenges: it is difficult to develop accurate performance prediction models using standard methods; the space of possible configurations is very large so exact solutions cannot be efficiently determined, and the mix and intensity of query classes in a workload vary dynamically over time. We provide a formulation of the provisioning problem and then define a framework to solve the problem. Our framework contains a cost model to predict the cost of executing a workload on a configuration and a method of selecting configurations. The cost model balances resource costs and penalties from SLAs. The specific resource demands and frequencies are accounted for by queueing network models of the Virtual Machines (VMs), which are used to predict performance. We evaluate our approach experimentally using sample data analytic workloads on Amazon EC2.
international symposium on computers and communications | 2011
Patrick Martin; Andrew Brown; Wendy Powley; José Luis Vázquez-Poletti
Cloud computing, with its support for elastic resources that are available on an on-demand, pay-as-you-go basis, is an attractive platform for hosting Web-based services that have variable demand, yet consistent performance requirements. Effective service management is mandatory in order for services running in the cloud, which we call elastic services, to be cost-effective. In this paper we describe a management framework to facilitate elasticity of resource consumption by services in the cloud. We extend our framework for services management with the necessary concepts and properties to support elastic services. A prototype implementation is described.
grid computing | 2007
José Luis Vázquez-Poletti; Eduardo Huedo; Rubén S. Montero; Ignacio Martín Llorente
In order to achieve a reasonable degree of performance and reliability, Metascheduling has been revealed as a key functionality of the grid middleware. The aim of this paper is to provide a comparative analysis between two major grid scheduling philosophies: a semi-centralized approach, represented by the EGEE Workload Management System, and a fully distributed approach, represented by the Grid Way Metascheduler. The distributed approach follow a loosely-coupled philosophy for the Grid resembling the end-to-end principle, which has fostered the spectacular development and diffusion of the Internet and, in particular, Web technologies in the past decade. The comparative is both theoretical, through a functionality checklist, and experimental, through the execution of a fusion physics plasma application on the EGEE infrastructure. This paper not only includes a standard analysis with the obtained times, but also a complex analysis based on a performance model.
intelligent data acquisition and advanced computing systems: technology and applications | 2013
Richard M. Wallace; Volodymyr Turchenko; Mehdi Sheikhalishahi; Iryna V. Turchenko; Vladyslav Shults; José Luis Vázquez-Poletti; Lucio Grandinetti
Advances in service-oriented architectures (SOA), virtualization, high-speed networks, and cloud computing has resulted in attractive pay-as-you-go services. Job scheduling on these systems results in commodity bidding for computing time. This bidding is institutionalized by Amazon for its Elastic Cloud Computing (EC2) environment and bidding methods exist for other cloud-computing vendors as well as multi-cloud and cluster computing brokers such as SpotCloud. Commodity bidding for computing has resulted in complex spot price models that have ad-hoc strategies to provide demand for excess capacity. In this paper we will discuss vendors who provide spot pricing and bidding and present a predictive model for future spot prices based on neural networking giving users a high confidence on future prices aiding bidding on commodity computing.
Future Generation Computer Systems | 2016
Mehdi Sheikhalishahi; Richard M. Wallace; Lucio Grandinetti; José Luis Vázquez-Poletti; Francesca Guerriero
With the advent of new computing technologies, such as cloud computing and contemporary parallel processing systems, the building blocks of computing systems have become multi-dimensional. Traditional scheduling systems based on a single-resource optimization, like processors, fail to provide near optimal solutions. The efficient use of new computing systems depends on the efficient use of several resource dimensions. Thus, the scheduling systems have to fully use all resources. In this paper, we address the problem of multi-resource scheduling via multi-capacity bin-packing. We propose the application of multi-capacity-aware resource scheduling at host selection layer and queuing mechanism layer of a scheduling system. The experimental results demonstrate performance improvements of scheduling in terms of waittime and slowdown metrics. A proposal for scheduling problem based on multi-capacity bin-packing algorithms.A proposal for host selection and queuing based on multi-resource scheduling.Getting better waittime and slowdown metrics than the state of the art scheduling.
Software - Practice and Experience | 2015
Mehdi Sheikhalishahi; Lucio Grandinetti; Richard M. Wallace; José Luis Vázquez-Poletti
The complexity of computing systems introduces a few issues and challenges such as poor performance and high energy consumption. In this paper, we first define and model resource contention metric for high performance computing workloads as a performance metric in scheduling algorithms and systems at the highest level of resource management stack to address the main issues in computing systems. Second, we propose a novel autonomic resource contention‐aware scheduling approach architected on various layers of the resource management stack. We establish the relationship between distributed resource management layers in order to optimize resource contention metric. The simulation results confirm the novelty of our approach.Copyright
Concurrency and Computation: Practice and Experience | 2014
Ginés D. Guerrero; Richard M. Wallace; José Luis Vázquez-Poletti; José M. Cecilia; José M. García; Daniel Mozos; Horacio Pérez-Sánchez
Virtual Screening (VS) methods can considerably aid drug discovery research, predicting how ligands interact with drug targets. BINDSURF is an efficient and fast blind VS methodology for the determination of protein binding sites, depending on the ligand, using the massively parallel architecture of graphics processing units(GPUs) for fast unbiased prescreening of large ligand databases. In this contribution, we provide a performance/cost model for the execution of this application on both local system and public cloud infrastructures. With our model, it is possible to determine which is the best infrastructure to use in terms of execution time and costs for any given problem to be solved by BINDSURF. Conclusions obtained from our study can be extrapolated to other GPU‐based VS methodologies.Copyright
international conference on parallel processing | 2015
Rui Han; Junwei Wang; Siguang Huang; Chenrong Shao; Shulin Zhan; Jianfeng Zhan; José Luis Vázquez-Poletti
Modern latency-critical online services often rely on composing results from a large number of server components. Hence the tail latency (e.g. The 99th percentile of response time), rather than the average, of these components determines the overall service performance. When hosted on a cloud environment, the components of a service typically co-locate with short batch jobs to increase machine utilizations, and share and contend resources such as caches and I/O bandwidths with them. The highly dynamic nature of batch jobs in terms of their workload types and input sizes causes continuously changing performance interference to individual components, hence leading to their latency variability and high tail latency. However, existing techniques either ignore such fine-grained component latency variability when managing service performance, or rely on executing redundant requests to reduce the tail latency, which adversely deteriorate the service performance when load gets heavier. In this paper, we propose PCS, a predictive and component-level scheduling framework to reduce tail latency for large-scale, parallel online services. It uses an analytical performance model to simultaneously predict the component latency and the overall service performance on different nodes. Based on the predicted performance, the scheduler identifies straggling components and conducts near-optimal component-node allocations to adapt to the changing performance interferences from batch jobs. We demonstrate that, using realistic workloads, the proposed scheduler reduces the component tail latency by an average of 67.05% and the average overall service latency by 64.16% compared with the state-of-the-art techniques on reducing tail latency.
cluster computing and the grid | 2007
José Luis Vázquez-Poletti; Eduardo Huedo; Rubén S. Montero; Ignacio Martín Llorente
Bioinformatics is demanding more computational resources day after day. The problems proposed by this area are growing in such complexity that traditional computing systems are not able to face them. For solving complex problems which can be divided in tasks with dependencies, a workflow management system must be employed. In this paper, we introduce the use of the workflow management of the GridWay metascheduler for running a Bioinformatics application which implements a complex algorithm performing protein clustering in order to obtain non-redundant protein databases. The use of a general purpose meta-scheduling system will provide the application the fault-tolerance and advance scheduling capabilities needed to execute on a highly dynamic, heterogeneous and faulty environment. The execution results on a production Grid (the EGEE infrastructure) shows the dramatic impact of remote queue waiting times on the application performance; and the critical need of efficient re-scheduling capabilities.