Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Samuel Cadellin Skipsey is active.

Publication


Featured researches published by Samuel Cadellin Skipsey.


Journal of Physics: Conference Series | 2012

Multi-core job submission and grid resource scheduling for ATLAS AthenaMP

D. R. M. Crooks; P Calafiura; R Harrington; M Jha; T Maeno; S Purdie; H Severini; Samuel Cadellin Skipsey; V Tsulaia; R Walker; A. Washbrook

AthenaMP is the multi-core implementation of the ATLAS software framework and allows the efficient sharing of memory pages between multiple threads of execution. This has now been validated for production and delivers a significant reduction on the overall application memory footprint with negligible CPU overhead. Before AthenaMP can be routinely run on the LHC Computing Grid it must be determined how the computing resources available to ATLAS can best exploit the notable improvements delivered by switching to this multi-process model. A study into the effectiveness and scalability of AthenaMP in a production environment will be presented. Best practices for configuring the main LRMS implementations currently used by grid sites will be identified in the context of multi-core scheduling optimisation.


Journal of Physics: Conference Series | 2015

Evaluation of containers as a virtualisation alternative for HEP workloads

Gareth Roy; A. Washbrook; D. R. M. Crooks; Gang Qin; Samuel Cadellin Skipsey; Gordon Stewart; David Britton

In this paper the emerging technology of Linux containers is examined and evaluated for use in the High Energy Physics (HEP) community. Key technologies required to enable containerisation will be discussed along with emerging technologies used to manage container images. An evaluation of the requirements for containers within HEP will be made and benchmarking will be carried out to asses performance over a range of HEP workflows. The use of containers will be placed in a broader context and recommendations on future work will be given.


Journal of Physics: Conference Series | 2014

Monitoring in a grid cluster

D. R. M. Crooks; Mark Mitchell; Stuart Purdie; Gareth Roy; Samuel Cadellin Skipsey; David Britton

The monitoring of a grid cluster (or of any piece of reasonably scaled IT infrastructure) is a key element in the robust and consistent running of that site. There are several factors which are important to the selection of a useful monitoring framework, which include ease of use, reliability, data input and output. It is critical that data can be drawn from different instrumentation packages and collected in the framework to allow for a uniform view of the running of a site. It is also very useful to allow different views and transformations of this data to allow its manipulation for different purposes, perhaps unknown at the initial time of installation. In this context, we present the findings of an investigation of the Graphite monitoring framework and its use at the ScotGrid Glasgow site. In particular, we examine the messaging system used by the framework and means to extract data from different tools, including the existing framework Ganglia which is in use at many sites, in addition to adapting and parsing data streams from external monitoring frameworks and websites.


arXiv: Distributed, Parallel, and Cluster Computing | 2011

Establishing Applicability of SSDs to LHC Tier-2 Hardware Configuration

Samuel Cadellin Skipsey; W. Bhimji; M. Kenyon

Solid State Disk technologies are increasingly replacing high-speed hard disks as the storage technology in high-random-I/O environments. There are several potentially I/O bound services within the typical LHC Tier-2 - in the back-end, with the trend towards many-core architectures continuing, worker nodes running many single-threaded jobs and storage nodes delivering many simultaneous files can both exhibit I/O limited efficiency. We estimate the effectiveness of affordable SSDs in the context of worker nodes, on a large Tier-2 production setup using both low level tools and real LHC I/O intensive data analysis jobs comparing and contrasting with high performance spinning disk based solutions. We consider the applicability of each solution in the context of its price/performance metrics, with an eye on the pragmatic issues facing Tier-2 provision and upgrades.


Journal of Physics: Conference Series | 2016

Cluster Optimisation using Cgroups at a Tier-2

G. Qin; Gareth Roy; D. R. M. Crooks; Samuel Cadellin Skipsey; Gordon Stewart; David Britton

The Linux kernel feature Control Groups (cgroups) has been used to gather metrics on the resource usage of single and eight-core ATLAS workloads. It has been used to study the effects on performance of a reduction in the amount of physical memory. The results were used to optimise cluster performance, and consequently increase cluster throughput by up to 10%.


arXiv: Distributed, Parallel, and Cluster Computing | 2015

Extending DIRAC File Management with Erasure-Coding for efficient storage.

Samuel Cadellin Skipsey; Paulin Todev; David Britton; D. R. M. Crooks; Gareth Roy

The state of the art in Grid style data management is to achieve increased resilience of data via multiple complete replicas of data files across multiple storage endpoints. While this is effective, it is not the most space-efficient approach to resilience, especially when the reliability of individual storage endpoints is sufficiently high that only a few will be inactive at any point in time. We report on work performed as part of GridPP\cite{GridPP}, extending the Dirac File Catalogue and file management interface to allow the placement of erasure-coded files: each file distributed as N identically-sized chunks of data striped across a vector of storage endpoints, encoded such that any M chunks can be lost and the original file can be reconstructed. The tools developed are transparent to the user, and, as well as allowing up and downloading of data to Grid storage, also provide the possibility of parallelising access across all of the distributed chunks at once, improving data transfer and IO performance. We expect this approach to be of most interest to smaller VOs, who have tighter bounds on the storage available to them, but larger (WLCG) VOs may be interested as their total data increases during Run 2. We provide an analysis of the costs and benefits of the approach, along with future development and implementation plans in this area. In general, overheads for multiple file transfers provide the largest issue for competitiveness of this approach at present.


Journal of Physics: Conference Series | 2015

Development of site-oriented Analytics for Grid computing centres

A. Washbrook; D. R. M. Crooks; Gareth Roy; Samuel Cadellin Skipsey; G. Qin; Gordon Stewart; David Britton

The field of analytics, the process of analysing data to visualise meaningful patterns and trends, has become increasingly important in scientific computing as the volume and variety of data available to process has significantly increased. There is now ongoing work in the High Energy Physics (HEP) community in this area, for example in the augmentation of systems management at WLCG computing sites. We report on work evaluating the feasibility of distributed site-oriented analytics using the Elasticsearch, Logstash and Kibana software stack and demonstrate functionality by the application of two workflows that give greater insight into site operations.


Journal of Physics: Conference Series | 2012

Analysing I/O bottlenecks in LHC data analysis on grid storage resources

W. Bhimji; P J Clark; M Doidge; M P Hellmich; Samuel Cadellin Skipsey; I Vukotic

We describe recent I/O testing frameworks that we have developed and applied within the UK GridPP Collaboration, the ATLAS experiment and the DPM team, for a variety of distinct purposes. These include benchmarking vendor supplied storage products, discovering scaling limits of SRM solutions, tuning of storage systems for experiment data analysis, evaluating file access protocols, and exploring I/O read patterns of experiment software and their underlying event data models. With multiple grid sites now dealing with petabytes of data, such studies are becoming essential. We describe how the tests build, and improve, on previous work and contrast how the use-cases differ. We also detail the results obtained and the implications for storage hardware, middleware and experiment software.


Journal of Physics: Conference Series | 2012

Taking the C out of CVMFS

Samuel Cadellin Skipsey; T. Doherty; Andy Turner

The Cern Virtual Machine File System is most well known as a distribution mechanism for the WLCG VOs@@ experiment software; as a result, almost all the existing expertise is in installing clients mount the central Cern repositories. We report the results of an initial experiment in using the cvmfs server packages to provide Glasgow-based repository aimed at software provisioning for small UK-local VOs. In general, although the documentation is sparse, server configuration is reasonably easy, with some experimentation. We discuss the advantages of local CVMFS repositories for sites, with some examples from our test VOs, vo.optics.ac.uk and neiss.org.uk.


Journal of Physics: Conference Series | 2012

Testing performance of standards-based protocols in DPM

Samuel Cadellin Skipsey; W. Bhimji; Ricardo Rocha

In the interests of the promotion of the increased use of non-proprietary protocols in grid storage systems, we perform tests on the performance of WebDAV and pNFS transport with the DPM storage solution. We find that the standards-based protocols behave similarly to the proprietary standards currently in use, despite encountering some issues with the state of the implementation itself. We thus conclude that there is no performance-based reason to avoid using such protocols for data management in future.

Collaboration


Dive into the Samuel Cadellin Skipsey's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

W. Bhimji

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Stuart Purdie

University of St Andrews

View shared research outputs
Top Co-Authors

Avatar

A. Washbrook

University of Edinburgh

View shared research outputs
Top Co-Authors

Avatar

M. Kenyon

University of Glasgow

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge