Jorda Polo
Polytechnic University of Catalonia
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Jorda Polo.
network operations and management symposium | 2010
Jorda Polo; David Carrera; Yolanda Becerra; Malgorzata Steinder; Ian Whalley
MapReduce is a data-driven programming model proposed by Google in 2004 which is especially well suited for distributed data analytics applications. We consider the management of MapReduce applications in an environment where multiple applications share the same physical resources. Such sharing is in line with recent trends in data center management which aim to consolidate workloads in order to achieve cost and energy savings. In a shared environment, it is necessary to predict and manage the performance of workloads given a set of performance goals defined for them. In this paper, we address this problem by introducing a new task scheduler for a MapReduce framework that allows performance-driven management of MapReduce tasks. The proposed task scheduler dynamically predicts the performance of concurrent MapReduce jobs and adjusts the resource allocation for the jobs. It allows applications to meet their performance objectives without over-provisioning of physical resources.
international middleware conference | 2011
Jorda Polo; Claris Castillo; David Carrera; Yolanda Becerra; Ian Whalley; Malgorzata Steinder; Jordi Torres; Eduard Ayguadé
We present a resource-aware scheduling technique for MapReduce multi-job workloads that aims at improving resource utilization across machines while observing completion time goals. Existing MapReduce schedulers define a static number of slots to represent the capacity of a cluster, creating a fixed number of execution slots per machine. This abstraction works for homogeneous workloads, but fails to capture the different resource requirements of individual jobs in multi-user environments. Our technique leverages job profiling information to dynamically adjust the number of slots on each machine, as well as workload placement across them, to maximize the resource utilization of the cluster. In addition, our technique is guided by user-provided completion time goals for each job. Source code of our prototype is available at [1].
IEEE Transactions on Network and Service Management | 2013
Jorda Polo; Yolanda Becerra; David Carrera; Malgorzata Steinder; Ian Whalley; Jordi Torres; Eduard Ayguadé
This paper presents a scheduling technique for multi-job MapReduce workloads that is able to dynamically build performance models of the executing workloads, and then use these models for scheduling purposes. This ability is leveraged to adaptively manage workload performance while observing and taking advantage of the particulars of the execution environment of modern data analytics applications, such as hardware heterogeneity and distributed storage. The technique targets a highly dynamic environment in which new jobs can be submitted at any time, and in which MapReduce workloads share physical resources with other workloads. Thus the actual amount of resources available for applications can vary over time. Beyond the formulation of the problem and the description of the algorithm and technique, a working prototype (called Adaptive Scheduler) has been implemented. Using the prototype and medium-sized clusters (of the order of tens of nodes), the following aspects have been studied separately: the schedulers ability to meet high-level performance goals guided only by user-defined completion time goals; the schedulers ability to favor data-locality in the scheduling algorithm; and the schedulers ability to deal with hardware heterogeneity, which introduces hardware affinity and relative performance characterization for those applications that can benefit from executing on specialized processors.
network computing and applications | 2015
Marcelo Amaral; Jorda Polo; David Carrera; Iqbal Mohomed; Merve Unuvar; Malgorzata Steinder
Micro services architecture has started a new trend for application development for a number of reasons: (1) to reduce complexity by using tiny services, (2) to scale, remove and deploy parts of the system easily, (3) to improve flexibility to use different frameworks and tools, (4) to increase the overall scalability, and (5) to improve the resilience of the system. Containers have empowered the usage of micro services architectures by being lightweight, providing fast start-up times, and having a low overhead. Containers can be used to develop applications based on monolithic architectures where the whole system runs inside a single container or inside a micro services architecture where one or few processes run inside the containers. Two models can be used to implement a micro services architecture using containers: master-slave, or nested-container. The goal of this work is to compare the performance of CPU and network running benchmarks in the two aforementioned models of micro services architecture hence provide a benchmark analysis guidance for system designers.
cluster computing and the grid | 2014
Jorda Polo; Yolanda Becerra; David Carrera; Jordi Torres; Eduard Ayguadé; Malgorzata Steinder
In this paper we present a MapReduce task scheduler for shared environments in which MapReduce is executed along with other resource-consuming workloads, such as transactional applications. All workloads may potentially share the same data store, some of them consuming data for analytics purposes while others acting as data generators. This kind of scenario is becoming increasingly important in data centers where improved resource utilization can be achieved through workload consolidation, and is specially challenging due to the interaction between workloads of different nature that compete for limited resources. The proposed scheduler aims to improve resource utilization across machines while observing completion time goals. Unlike other MapReduce schedulers, our approach also takes into account the resource demands for non-MapReduce workloads, and assumes that the amount of resources made available to the MapReduce applications is variable over time. As shown in our experiments, our proposal improves the management of MapReduce jobs in the presence of variable resource availability, increasing the accuracy of the estimations made by the scheduler, thus improving completion time goals without an impact on the fairness of the scheduler.
network computing and applications | 2013
Jorda Polo; Yolanda Becerra; David Carrera; Jordi Torres; Eduard Ayguadé; Mike Spreitzer; Malgorzata Steinder
Current distributed key-value stores generally provide greater scalability at the expense of weaker consistency and isolation. However, additional isolation support is becoming increasingly important in the environments in which these stores are deployed, where different kinds of applications with different needs are executed, from transactional workloads to data analytics. While fully-fledged ACID support may not be feasible, it is still possible to take advantage of the design of these data stores, which often include the notion of multiversion concurrency control, to enable them with additional features at a much lower performance cost and maintaining its scalability and availability. In this paper we explore the effects that additional consistency guarantees and isolation capabilities may have on a state of the art key-value store: Apache Cassandra. We propose and implement a new multiversioned isolation level that provides stronger guarantees without compromising Cassandras scalability and availability. As shown in our experiments, our version of Cassandra allows Snapshot Isolation-like transactions, preserving the overall performance and scalability of the system.
ieee international conference on high performance computing data and analytics | 2017
Marcelo Amaral; Jorda Polo; David Carrera; Seetharami R. Seelam; Malgorzata Steinder
Recent advances in hardware, such as systems with multiple GPUs and their availability in the cloud, are enabling deep learning in various domains including health care, autonomous vehicles, and Internet of Things. Multi-GPU systems exhibit complex connectivity among GPUs and between GPUs and CPUs. Workload schedulers must consider hardware topology and workload communication requirements in order to allocate CPU and GPU resources for optimal execution time and improved utilization in shared cloud environments. This paper presents a new topology-aware workload placement strategy to schedule deep learning jobs on multi-GPU systems. The placement strategy is evaluated with a prototype on a Power8 machine with Tesla P100 cards, showing speedups of up to ≈1.30x compared to state-of-the-art strategies; the proposed algorithm achieves this result by allocating GPUs that satisfy workload requirements while preventing interference. Additionally, a large-scale simulation shows that the proposed strategy provides higher resource utilization and performance in cloud systems.
international bhurban conference on applied sciences and technology | 2016
Muhammad Shafiq; Jorda Polo; Branimir Dickov; Tassadaq Hussain
Smith-Waterman algorithm is primarily used in DNA and protein sequencing which helps by a local sequence alignment to determine similarities between biomolecule sequences. However the inefficiency in performance of this algorithm limits its applications in the real world. In this perspective, this work presents two fold contributions. It develops and evaluates a mathematical performance model for the algorithm by targeting a distributed processing system. This mathematical model can be helpful to estimate performance of the algorithm for larger size of sequences aligned by the thread level parallelism, using large set of processors configured as distributed processing nodes. Secondly, This work also evaluates in detail the performance scalability of smith-waterman algorithm using OpenMP, MP and a Hybrid (OpenMP + MPI) parallel programming models on a real supercomputing platform Altix-4700. This evaluation shows that the hybrid approach performs better than the other simple approaches.
international conference on parallel processing | 2010
Jorda Polo; David Carrera; Yolanda Becerra; Vicenç Beltran; Jordi Torres; Eduard Ayguadé
international conference on big data | 2018
Shuja-ur-Rehman Baig; Marcelo Amaral; Jorda Polo; David Carrera