Katherine Riley
Argonne National Laboratory
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Katherine Riley.
parallel computing | 2009
Anshu Dubey; Katie Antypas; Murali K. Ganapathy; Lynn B. Reid; Katherine Riley; Daniel J. Sheeler; Andrew R. Siegel; Klaus Weide
FLASH is a publicly available high performance application code which has evolved into a modular, extensible software system from a collection of unconnected legacy codes. FLASH has been successful because its capabilities have been driven by the needs of scientific applications, without compromising maintainability, performance, and usability. In its newest incarnation, FLASH3 consists of inter-operable modules that can be combined to generate different applications. The FLASH architecture allows arbitrarily many alternative implementations of its components to co-exist and interchange with each other, resulting in greater flexibility. Further, a simple and elegant mechanism exists for customization of code functionality without the need to modify the core implementation of the source. A built-in unit test framework providing verifiability, combined with a rigorous software maintenance process, allow the code to operate simultaneously in the dual mode of production and development. In this paper we describe the FLASH3 architecture, with emphasis on solutions to the more challenging conflicts arising from solver complexity, portable performance requirements, and legacy codes. We also include results from user surveys conducted in 2005 and 2007, which highlight the success of the code.
international conference on cluster computing | 2009
Philip H. Carns; Robert Latham; Robert B. Ross; Kamil Iskra; Samuel Lang; Katherine Riley
Developing and tuning computational science applications to run on extreme scale systems are increasingly complicated processes. Challenges such as managing memory access and tuning message-passing behavior are made easier by tools designed specifically to aid in these processes. Tools that can help users better understand the behavior of their application with respect to I/O have not yet reached the level of utility necessary to play a central role in application development and tuning. This deficiency in the tool set means that we have a poor understanding of how specific applications interact with storage. Worse, the community has little knowledge of what sorts of access patterns are common in todays applications, leading to confusion in the storage research community as to the pressing needs of the computational science community. This paper describes the Darshan I/O characterization tool. Darshan is designed to capture an accurate picture of application I/O behavior, including properties such as patterns of access within files, with the minimum possible overhead. This characterization can shed important light on the I/O behavior of applications at extreme scale. Darshan also can enable researchers to gain greater insight into the overall patterns of access exhibited by such applications, helping the storage community to understand how to best serve current computational science applications and better predict the needs of future applications. In this work we demonstrate Darshans ability to characterize the I/O behavior of four scientific applications and show that it induces negligible overhead for I/O intensive jobs with as many as 65,536 processes.
european conference on parallel processing | 2005
George S. Almasi; Gyan Bhanot; Dong Chen; Maria Eleftheriou; Blake G. Fitch; Alan Gara; Robert S. Germain; John A. Gunnels; Manish Gupta; Philip Heidelberg; Mike Pitman; Alek sandr Rayshubskiy; James C. Sexton; Frank Suits; Pavlos M. Vranas; Bob Walkup; Christopher Ward; Yuriy Zhestkov; Alessandro Curioni; Wanda Andreoni; Charles J. Archer; José E. Moreira; Richard D. Loft; Henry M. Tufo; Theron Voran; Katherine Riley
Blue Gene/L uses a large number of low power processors, together with multiple integrated interconnection networks, to build a supercomputer with low cost, space and power consumption. It uses a novel system software architecture designed with application scalability in mind. However, whether real applications will scale to tens of thousands of processors has been an open question. In this paper, we describe early experience with several applications on a 16,384 node Blue Gene/L system. This study establishes that applications from a broad variety of scientific disciplines can effectively scale to thousands of processors. The results reported in this study represent the highest performance ever demonstrated for most of these applications, and in fact, show effective scaling for the first time ever on thousands of processors.
IEEE Computer | 2015
Nichols A. Romero; Aiichiro Nakano; Katherine Riley; Fuyuki Shimojo; Rajiv K. Kalia; Priya Vashishta; Paul Messina
As the scale of quantum molecular dynamics simulations has grown in time and system size, QMD codes must increase intranode and instruction-level parallelism to take advantage of emerging supercomputer architectures. The authors present one promising parallelization approach and illustrate its success on one of the worlds most powerful systems.
ieee international conference on high performance computing data and analytics | 2014
Anshu Dubey; Katie Antypas; Alan Clark Calder; Christopher S. Daley; Bruce Fryxell; Brad Gallagher; Donald Q. Lamb; Dongwook Lee; Kevin Olson; Lynn B. Reid; Paul Rich; Paul M. Ricker; Katherine Riley; R. Rosner; Andrew R. Siegel; Noel T. Taylor; Klaus Weide; Francis Xavier Timmes; Natasha Vladimirova; John A. ZuHone
The FLASH code has evolved into a modular and extensible scientific simulation software system over the decade of its existence. During this time it has been cumulatively used by over a thousand researchers to investigate problems in astrophysics, cosmology, and in some areas of basic physics, such as turbulence. Recently, many new capabilities have been added to the code to enable it to simulate problems in high-energy density physics. Enhancements to these capabilities continue, along with enhancements enabling simulations of problems in fluid-structure interactions. The code started its life as an amalgamation of already existing software packages and sections of codes developed independently by various participating members of the team for other purposes. The code has evolved through a mixture of incremental and deep infrastructural changes. In the process, it has undergone four major revisions, three of which involved a significant architectural advancement. Along the way, a software process evolved that addresses the issues of code verification, maintainability, and support for the expanding user base. The software process also resolves the conflicts arising out of being in development and production simultaneously with multiple research projects, and between performance and portability. This paper describes the process of code evolution with emphasis on the design decisions and software management policies that have been instrumental in the success of the code. The paper also makes the case for a symbiotic relationship between scientific research and good software engineering of the simulation software.
computational science and engineering | 2013
Anshu Dubey; Katie Antypas; Alan Clark Calder; Bruce Fryxell; D. Q. Lamb; Paul M. Ricker; Lynn B. Reid; Katherine Riley; R. Rosner; Andrew R. Siegel; F. X. Timmes; Natalia Vladimirova; Klaus Weide
The FLASH code has evolved into a modular and extensible scientific simulation software system over the decade of its existence. During this time it has been cumulatively used by over a thousand researchers in several scientific communities (i.e. astrophysics, cosmology, high-energy density physics, turbulence, fluid-structure interactions) to obtain results for research. The code started its life as an amalgamation of two already existing software packages and sections of other codes developed independently by various participating members of the team for other purposes. In the evolution process it has undergone four major revisions, three of which involved a significant architectural advancement. A corresponding evolution of the software process and policies for maintenance occurred simultaneously. The code is currently in its 4.x release with a substantial user community. Recently there has been an upsurge in the contributions by external users; some provide significant new capability. This paper outlines the software development and evolution processes that have contributed to the success of the FLASH code.
Ibm Journal of Research and Development | 2013
Susan Coghlan; Kalyan Kumaran; Raymond M. Loy; Paul Messina; Vitali A. Morozov; James C. Osborn; Scott Parker; Katherine Riley; Nichols A. Romero; Timothy J. Williams
A varied collection of scientific and engineering codes has been adapted and enhanced to take advantage of the IBM Blue Gene®/Q architecture and thus enable research that was previously out of reach. Computational research teams from a number of disciplines collaborated with the staff of the Argonne Leadership Computing Facility to assess which of Blue Gene/Qs many novel features could be exploited for each application to equip it to tackle existing problem classes with greater fidelity and in some cases to address new phenomena. The quad floating-point units and the five-dimensional torus interconnect are among the features that were demonstrated to be effective for a number of important applications. Furthermore, data obtained from the hardware counters provided insights that were valuable in guiding the code modifications. Hardware features and programming techniques that were effective across multiple codes are documented as well. First, we have confirmed that there is no significant code rewrite needed to run todays production codes with good performance on Mira, an IBM Blue Gene/Q supercomputer. Performance improvements are already demonstrated, even though our measurements are all on pre-production software and hardware. The application domains included biology, materials science, combustion, chemistry, nuclear physics, and industrial-scale design of nuclear reactors, jet engines, and the efficiency of transportation systems.
Archive | 2018
Richard A. Gerber; James J. Hack; Katherine Riley; Katie Antypas; Richard M. Coffey; Eli Dart; Tjerk Straatsma; J. C. Wells; Deborah Bard; Sudip S. Dosanjh; Inder Monga; Michael E. Papka; Lauren Rotman
Archive | 2015
Aurora Clark; Andy Millis; Laura Gagliardi; Thanos Panagiotopoulos; Ilja Siepmann; Chris Wolverton; Priya Vashishta; Mark Stevens; Mark S. Gordon; Paul R. C. Kent; Kerstin Kleese va DAm; Thomas Proffen; Craig Tull; Lori Diachin; Jamie Sethian; Anouar Benali; Jackie Chen; Katie Antypas; Richard A. Gerber; Katherine Riley; Tjerk Straatsma
Archive | 2015
Salman Habib; R.M. Roser; Richard A. Gerber; Katie Antypas; Katherine Riley; Timothy H. Williams; J. C. Wells; Tjerk Straatsma