Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ian Gable is active.

Publication


Featured researches published by Ian Gable.


Journal of Physics: Conference Series | 2008

Deploying HEP applications using Xen and Globus Virtual Workspaces

A Agarwal; Ronald J. Desmarais; Ian Gable; D Grundy; D P-Brown; R Seuster; Daniel C. Vanderster; Andre Charbonneau; R Enge; Randall Sobie

The deployment of HEP applications in heterogeneous grid environments can be challenging because many of the applications are dependent on specific OS versions and have a large number of complex software dependencies. Virtual machine monitors such as Xen could be used to package HEP applications, complete with their execution environments, to run on resources that do not meet their operating system requirements. Our previous work has shown HEP applications running within Xen suffer little or no performance penalty as a result of virtualization. However, a practical strategy is required for remotely deploying, booting, and controlling virtual machines on a remote cluster. One tool that promises to overcome the deployment hurdles using standard grid technology is the Globus Virtual Workspaces project. We describe strategies for the deployment of Xen virtual machines using Globus Virtual Workspace middleware that simplify the deployment of HEP applications.


scientific cloud computing | 2013

HTC scientific computing in a distributed cloud environment

Randall Sobie; A Agarwal; Ian Gable; Colin Leavett-Brown; Michael Paterson; Ryan Taylor; Andre Charbonneau; Roger Impey; Wayne Podiama

This paper describes the use of a distributed cloud computing system for high-throughput computing (HTC) scientific applications. The distributed cloud computing system is composed of a number of separate Infrastructure-as-a-Service (IaaS) clouds that are utilized in a unified infrastructure. The distributed cloud has been in production-quality operation for two years with approximately 500,000 completed jobs where a typical workload has 500 simultaneous embarrassingly-parallel jobs that run for approximately 12 hours. We review the design and implementation of the system which is based on pre-existing components and a number of custom components. We discuss the operation of the system, and describe our plans for the expansion to more sites and increased computing capacity.


Journal of Physics: Conference Series | 2011

A batch system for HEP applications on a distributed IaaS cloud

Ian Gable; A Agarwal; M Anderson; Patrick Armstrong; K Fransham; D Harris C Leavett-Brown; M Paterson; D Penfold-Brown; Randall Sobie; M Vliet; Andre Charbonneau; Roger Impey; Wayne Podaima

The emergence of academic and commercial Infrastructure-as-a-Service (IaaS) clouds is opening access to new resources for the HEP community. In this paper we will describe a system we have developed for creating a single dynamic batch environment spanning multiple IaaS clouds of different types (e.g. Nimbus, OpenNebula, Amazon EC2). A HEP user interacting with the system submits a job description file with a pointer to their VM image. VM images can either be created by users directly or provided to the users. We have created a new software component called Cloud Scheduler that detects waiting jobs and boots the user VM required on any one of the available cloud resources. As the user VMs appear, they are attached to the job queues of a central Condor job scheduler, the job scheduler then submits the jobs to the VMs. The number of VMs available to the user is expanded and contracted dynamically depending on the number of user jobs. We present the motivation and design of the system with particular emphasis on Cloud Scheduler. We show that the system provides the ability to exploit academic and commercial cloud sites in a transparent fashion.


Journal of Physics: Conference Series | 2010

A comparison of HEP code with SPEC1 benchmarks on multi-core worker nodes

Michele Michelotto; Manfred Alef; Alejandro Iribarren; H. Meinhard; Peter Wegner; Martin Bly; G. Benelli; Franco Brasolin; Hubert Degaudenzi; Alessandro De Salvo; Ian Gable; A. Hirstius; P. Hristov

The SPEC[1] CINT benchmark has been used as a performance reference for computing in the HEP community for the past 20 years. The SPECint_base2000 (SI2K) unit of performance has been used by the major HEP experiments both in the Computing Technical Design Report for the LHC experiments and in the evaluation of the Computing Centres. At recent HEPiX[3] meetings several HEP sites have reported disagreements between actual machine performances and the scores reported by SPEC. Our group performed a detailed comparison of Simulation and Reconstruction code performances from the four LHC experiments in order to find a successor to the SI2K benchmark. We analyzed the new benchmarks from SPEC CPU2006 suite, both integer and floating point, in order to find the best agreement with the HEP code behaviour, with particular attention paid to reproducing the actual environment of HEP farm i.e., each job running independently on each core, and matching compiler, optimization, percentage of integer and floating point operations, and ease of use.


Journal of Physics: Conference Series | 2010

Research computing in a distributed cloud environment

K Fransham; A Agarwal; Patrick Armstrong; A Bishop; Andre Charbonneau; Ronald J. Desmarais; N Hill; Ian Gable; S Gaudet; S Goliath; Roger Impey; Colin Leavett-Brown; J Ouellete; M Paterson; Chris Pritchet; D Penfold-Brown; Wayne Podaima; D Schade; Randall Sobie

The recent increase in availability of Infrastructure-as-a-Service (IaaS) computing clouds provides a new way for researchers to run complex scientific applications. However, using cloud resources for a large number of research jobs requires significant effort and expertise. Furthermore, running jobs on many different clouds presents even more difficulty. In order to make it easy for researchers to deploy scientific applications across many cloud resources, we have developed a virtual machine resource manager (Cloud Scheduler) for distributed compute clouds. In response to a users job submission to a batch system, the Cloud Scheduler manages the distribution and deployment of user-customized virtual machines across multiple clouds. We describe the motivation for and implementation of a distributed cloud using the Cloud Scheduler that is spread across both commercial and dedicated private sites, and present some early results of scientific data analysis using the system.


arXiv: Distributed, Parallel, and Cluster Computing | 2014

Dynamic web cache publishing for IaaS clouds using Shoal

Ian Gable; Michael Chester; Patrick Armstrong; F. Berghaus; Andre Charbonneau; Colin Leavett-Brown; Michael Paterson; Robert Prior; Randall Sobie; Ryan Taylor

We have developed a highly scalable application, called Shoal, for tracking and utilizing a distributed set of HTTP web caches. Squid servers advertise their existence to the Shoal server via AMQP messaging by running Shoal Agent. The Shoal server provides a simple REST interface that allows clients to determine their closest Squid cache. Our goal is to dynamically instantiate Squid caches on IaaS clouds in response to client demand. Shoal provides the VMs on IaaS clouds with the location of the nearest dynamically instantiated Squid Cache. In this paper, we describe the design and performance of Shoal.


Proceedings of SPIE | 2010

CANFAR: the Canadian Advanced Network for Astronomical Research

Severin J. Gaudet; Norman R. Hill; Patrick Armstrong; Nick Ball; Jeff Burke; Brian Chapel; Ed Chapin; Adrian Damian; Pat Dowler; Ian Gable; Sharon Goliath; Isabella Ghiurea; Sébastien Fabbro; Stephen Gwyn; Dustin Jenkins; J. J. Kavelaars; Brian Major; John Ouellette; M Paterson; Michael T. Peddle; Duncan Penfold-Brown; Chris Pritchet; David Schade; Randall Sobie; David Woods; Alinga Yeung; Yuehai Zhang

The Canadian Advanced Network For Astronomical Research (CANFAR) is a 2 1/2-year project that is delivering a network-enabled platform for the accessing, processing, storage, analysis, and distribution of very large astronomical datasets. The CANFAR infrastructure is being implemented as an International Virtual Observatory Alliance (IVOA) compliant web service infrastructure. A challenging feature of the project is to channel all survey data through Canadian research cyberinfrastructure. Sitting behind the portal service, the internal architecture makes use of high-speed networking, cloud computing, cloud storage, meta-scheduling, provisioning and virtualisation. This paper describes the high-level architecture and the current state of the project.


arXiv: Distributed, Parallel, and Cluster Computing | 2012

Data intensive high energy physics analysis in a distributed cloud

Andre Charbonneau; A Agarwal; M Anderson; Patrick Armstrong; K Fransham; Ian Gable; D Harris; Roger Impey; Colin Leavett-Brown; Michael Paterson; Wayne Podaima; Randall Sobie; M Vliet

We show that distributed Infrastructure-as-a-Service (IaaS) compute clouds can be effectively used for the analysis of high energy physics data. We have designed a distributed cloud system that works with any application using large input data sets requiring a high throughput computing environment. The system uses IaaS-enabled science and commercial clusters in Canada and the United States. We describe the process in which a user prepares an analysis virtual machine (VM) and submits batch jobs to a central scheduler. The system boots the user-specific VM on one of the IaaS clouds, runs the jobs and returns the output to the user. The user application accesses a central database for calibration data during the execution of the application. Similarly, the data is located in a central location and streamed by the running application. The system can easily run one hundred simultaneous jobs in an efficient manner and should scale to many hundreds and possibly thousands of user jobs.


Journal of Physics: Conference Series | 2008

BaBar MC production on the Canadian grid using a web services approach

A Agarwal; Patrick Armstrong; Ronald J. Desmarais; Ian Gable; S Popov; Simon Ramage; S Schaffer; C Sobie; Randall Sobie; T Sulivan; Daniel C. Vanderster; Gabriel Mateescu; Wayne Podaima; Andre Charbonneau; Roger Impey; M Viswanathan; Darcy Quesnel

The present paper highlights the approach used to design and implement a web services based BaBar Monte Carlo (MC) production grid using Globus Toolkit version 4. The grid integrates the resources of two clusters at the University of Victoria, using the ClassAd mechanism provided by the Condor-G metascheduler. Each cluster uses the Portable Batch System (PBS) as its local resource management system (LRMS). Resource brokering is provided by the Condor matchmaking process, whereby the job and resource attributes are expressed as ClassAds. The important features of the grid are automatic registering of resource ClassAds to the central registry, ClassAds extraction from the registry to the metascheduler for matchmaking, and the incorporation of input/output file staging. Web-based monitoring is employed to track the status of grid resources and the jobs for an efficient operation of the grid. The performance of this new grid for BaBar jobs, and the existing Canadian computational grid (GridX1) based on Globus Toolkit version 2 is found to be consistent.


Journal of Physics: Conference Series | 2015

Integrating network and transfer metrics to optimize transfer efficiency and experiment workflows

Shawn Patrick McKee; M. Babik; S. Campana; A. Di Girolamo; T. Wildish; J. Closier; S. Roiser; C. Grigoras; I. Vukotic; M. Salichos; K. De; Vincent Garonne; J.A.D. Cruz; A. Forti; C.J. Walker; D. Rand; A. De Salvo; E. Mazzoni; Ian Gable; F. Chollet; L. Caillat; F. Schaer; Hsin-Yen Chen; U. Tigerstedt; G. Duckeck; B. Hoeft; A. Petzold; F. Lopez; Jose Flix; S. Stancu

The Worldwide LHC Computing Grid relies on the network as a critical part of its infrastructure and therefore needs to guarantee effective network usage and prompt detection and resolution of any network issues, including connection failures, congestion, traffic routing, etc. The WLCG Network and Transfer Metrics project aims to integrate and combine all network-related monitoring data collected by the WLCG infrastructure. This includes FTS monitoring information, monitoring data from the XRootD federation, as well as results of the perfSONAR tests. The main challenge consists of further integrating and analyzing this information in order to allow the optimizing of data transfers and workload management systems of the LHC experiments. In this contribution, we present our activity in commissioning WLCG perfSONAR network and integrating network and transfer metrics: We motivate the need for the network performance monitoring, describe the main use cases of the LHC experiments as well as status and evolution in the areas of configuration and capacity management, datastore and analytics, including integration of transfer and network metrics and operations and support.

Collaboration


Dive into the Ian Gable's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

A Agarwal

University of Victoria

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Roger Impey

National Research Council

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Wayne Podaima

National Research Council

View shared research outputs
Top Co-Authors

Avatar

Ryan Taylor

University of Victoria

View shared research outputs
Researchain Logo
Decentralizing Knowledge