Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Peter van Gemmeren is active.

Publication


Featured researches published by Peter van Gemmeren.


21st International Conference on Computing in High Energy and Nuclear Physics, CHEP 2015 | 2015

Fine grained event processing on HPCs with the ATLAS Yoda system

P. Calafiura; K. De; Wen Guan; T. Maeno; P. Nilsson; Danila Oleynik; S. Panitkin; V. Tsulaia; Peter van Gemmeren; Torre Wenaus

High performance computing facilities present unique challenges and opportunities for HEP event processing. The massive scale of many HPC systems means that fractionally small utilization can yield large returns in processing throughput. Parallel applications which can dynamically and efficiently fill any scheduling opportunities the resource presents benefit both the facility (maximal utilization) and the (compute-limited) science. The ATLAS Yoda system provides this capability to HEP-like event processing applications by implementing event-level processing in an MPI-based master-client model that integrates seamlessly with the more broadly scoped ATLAS Event Service. Fine grained, event level work assignments are intelligently dispatched to parallel workers to sustain full utilization on all cores, with outputs streamed off to destination object stores in near real time with similarly fine granularity, such that processing can proceed until termination with full utilization. The system offers the efficiency and scheduling flexibility of preemption without requiring the application actually support or employ check-pointing. We will present the new Yoda system, its motivations, architecture, implementation, and applications in ATLAS data processing at several US HPC centers.


international conference on cluster computing | 2009

The event data store and I/O framework for the ATLAS experiment at the Large Hadron Collider

Peter van Gemmeren; D. Malon

The ATLAS experiment at the Large Hadron Collider, a collaboration of approximately 2600 particle physicists worldwide, will soon produce tens of petabytes of data annually, with an equivalent amount of simulated data. Even in advance of data taking, ATLAS has already accumulated many petabytes of event data from detector commissioning activities and simulation. This paper provides an overview of the design and implementation of the scientific data store that hosts ATLAS event data, and describes how the store is populated, organized, and typically accessed. Aspects of the event I/O framework and supporting infrastructure are also described. The paper further discusses some of the challenges that must be addressed for a scientific data store of this scope, scale, and complexity.


international conference on cluster computing | 2010

Supporting high-performance I/O at the petascale: The event data store for ATLAS at the LHC

Peter van Gemmeren; D. Malon

A key to delivering high-performance scientific data stores at multi-petabyte scales and beyond is to provide an infrastructure that allows scientists to select, read, and process only the data that they need. The ATLAS experiment at the Large Hadron Collider (LHC) has deployed an event data store and an I/O framework designed from their inception to support efficient selective reading at many scales, from event-level selections to partial retrieval of the complexly-structured event data objects themselves. In-file metadata is used to accomplish the bookkeeping needed to accompany these data reduction strategies. This paper describes features of the ATLAS I/O and event data store infrastructure that make efficient selective reading possible.


Journal of Physics: Conference Series | 2015

Running ATLAS workloads within massively parallel distributed applications using Athena Multi-Process framework (AthenaMP)

P. Calafiura; C. Leggett; R. Seuster; V. Tsulaia; Peter van Gemmeren

AthenaMP is a multi-process version of the ATLAS reconstruction, simulation and data analysis framework Athena. By leveraging Linux fork and copy-on-write mechanisms, it allows for sharing of memory pages between event processors running on the same compute node with little to no change in the application code. Originally targeted to optimize the memory footprint of reconstruction jobs, AthenaMP has demonstrated that it can reduce the memory usage of certain configurations of ATLAS production jobs by a factor of 2. AthenaMP has also evolved to become the parallel event-processing core of the recently developed ATLAS infrastructure for fine-grained event processing (Event Service) which allows the running of AthenaMP inside massively parallel distributed applications on hundreds of compute nodes simultaneously. We present the architecture of AthenaMP, various strategies implemented by AthenaMP for scheduling workload to worker processes (for example: Shared Event Queue and Shared Distributor of Event Tokens) and the usage of AthenaMP in the diversity of ATLAS event processing workloads on various computing resources: Grid, opportunistic resources and HPC.


Journal of Physics: Conference Series | 2008

Explicit state representation and the ATLAS event data model: theory and practice

M Nowak; D. Malon; Peter van Gemmeren; A. C. Schaffer; S. Snyder; S Binet; K. Cranmer

In anticipation of data taking, ATLAS has undertaken a program of work to develop an explicit state representation of the experiments complex transient event data model. This effort has provided both an opportunity to consider explicitly the structure, organization, and content of the ATLAS persistent event store before writing tens of petabytes of data (replacing simple streaming, which uses the persistent store as a core dump of transient memory), and a locus for support of event data model evolution, including significant refactoring, beyond the automatic schema evolution capabilities of underlying persistence technologies. ATLAS has encountered the need for such non-trivial schema evolution on several occasions already. This paper describes the state representation strategy (transient/persistent separation) and its implementation, including both the payoffs that ATLAS has seen (significant and sometimes surprising space and performance improvements, the extra layer notwithstanding, and extremely general schema evolution support) and the costs (additional and relatively pervasive additional infrastructure development and maintenance). The paper further discusses how these costs are mitigated, and how ATLAS is able to implement this strategy without losing the ability to take advantage of the (improving!) automatic schema evolution capabilities of underlying technology layers when appropriate. Implications of state representations for direct ROOT browsability, and current strategies for associating physics analysis views with such state representations, are also described.


Proceedings of 38th International Conference on High Energy Physics — PoS(ICHEP2016) | 2017

Managing Asynchronous Data in ATLAS's Concurrent Framework

John Baines; V. Tsulaia; P. Calafiura; J. Cranshaw; Peter van Gemmeren; D. Malon; T. Bold; C. Leggett; Benjamin Wynne; A. Dotti; Scott Snyder; Graeme Stewart; S. Farrell

In order to be able to make effective use of emerging hardware, where the amount of memory available to any CPU is rapidly decreasing as the core count continues to rise, ATLAS has begun a migration to a concurrent, multi-threaded software framework, known as AthenaMT. Significant progress has been made in implementing AthenaMT - we can currently run realistic Geant4 simulations on massively concurrent machines. The migration of realistic prototypes of reconstruction workflows is more difficult, given the large amount of legacy code and the complexity and challenges of reconstruction software. These types of workflows, however, are the types that will most benefit from the memory reduction features of a multi-threaded framework. One of the challenges that we will report on in this paper is the re-design and implementation of several key asynchronous technologies whose behaviour is radically different in a concurrent environment than in a serial one, namely the management of Conditions data and the Detector Description, and the handling of asynchronous notifications (such as FileOpen). Since asynchronous data, such as Conditions or detector alignments, has a lifetime different than that of event data, it cannot be kept in the Event Store. However, multiple instances of the data need to be simultaneously accessible, such that concurrent events that are, for example, processing conditions data from different validity intervals can be executed concurrently in an efficient manner with low memory overhead, and without multi-threaded conflicts.


Proceedings of 38th International Conference on High Energy Physics — PoS(ICHEP2016) | 2017

ATLAS Event Data Organization and I/O Framework Capabilities in Support of Heterogeneous Data Access and Processing Models

D. Malon; Peter van Gemmeren; J. Cranshaw; M Nowak

Choices in persistent data models and data organization have significant performance ramifica- tions for data-intensive scientific computing. In experimental high energy physics, organizing file-based event data for efficient per-attribute retrieval may improve the I/O performance of some physics analyses but hamper the performance of processing that requires full-event access. In-file data organization tuned for serial access by a single process may be less suitable for opportunis- tic sub-file-based processing on distributed computing resources. Unique I/O characteristics of high-performance computing platforms pose additional challenges. The ATLAS experiment at the Large Hadron Collider employs a flexible I/O framework and a suite of tools and techniques for persistent data organization to support an increasingly heterogeneous array of data access and processing models.


Journal of Physics: Conference Series | 2017

AthenaMT: upgrading the ATLAS software framework for the many-core world with multi-threading

C. Leggett; V. Tsulaia; P. Calafiura; John Baines; Peter van Gemmeren; D. Malon; T. Bold; Benjamin Wynne; Scott Snyder; Graeme Stewart; S. Farrell; E. Ritsch

ATLASs current software framework, Gaudi/Athena, has been very successful for the experiment in LHC Runs 1 and 2. However, its single threaded design has been recognized for some time to be increasingly problematic as CPUs have increased core counts and decreased available memory per core. Even the multi-process version of Athena, AthenaMP, will not scale to the range of architectures we expect to use beyond Run2. After concluding a rigorous requirements phase, where many design components were examined in detail, ATLAS has begun the migration to a new data-flow driven, multi-threaded framework, which enables the simultaneous processing of singleton, thread unsafe legacy Algorithms, cloned Algorithms that execute concurrently in their own threads with different Event contexts, and fully re-entrant, thread safe Algorithms. In this paper we report on the process of modifying the framework to safely process multiple concurrent events in different threads, which entails significant changes in the underlying handling of features such as event and time dependent data, asynchronous callbacks, metadata, integration with the online High Level Trigger for partial processing in certain regions of interest, concurrent I/O, as well as ensuring thread safety of core services. We also report on upgrading the framework to handle Algorithms that are fully re-entrant.


Archive | 2017

Shared I/O components for the ATLAS multi-processing framework

Peter van Gemmeren; V. Tsulaia; D. Malon; M Nowak


Journal of Physics: Conference Series | 2017

Production Experience with the ATLAS Event Service

Douglas Benjamin; V. Tsulaia; P. Calafiura; Wen Guan; John Taylor Childers; Peter van Gemmeren; T. Maeno; K. De; Torre Wenaus; P. Nilsson

Collaboration


Dive into the Peter van Gemmeren's collaboration.

Top Co-Authors

Avatar

D. Malon

Argonne National Laboratory

View shared research outputs
Top Co-Authors

Avatar

V. Tsulaia

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

P. Calafiura

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

C. Leggett

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

M Nowak

Brookhaven National Laboratory

View shared research outputs
Top Co-Authors

Avatar

J. Cranshaw

Argonne National Laboratory

View shared research outputs
Top Co-Authors

Avatar

K. De

University of Texas at Arlington

View shared research outputs
Top Co-Authors

Avatar

P. Nilsson

Brookhaven National Laboratory

View shared research outputs
Top Co-Authors

Avatar

S. Farrell

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Scott Snyder

Brookhaven National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge