Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Melissa Romanus is active.

Publication


Featured researches published by Melissa Romanus.


Journal of Chemical Theory and Computation | 2015

Characterization of the three-dimensional free energy manifold for the uracil ribonucleoside from asynchronous replica exchange simulations.

Brian K. Radak; Melissa Romanus; Tai-Sung Lee; Haoyuan Chen; Ming Huang; Antons Treikalis; Vivekanandan Balasubramanian; Shantenu Jha; Darrin M. York

Replica exchange molecular dynamics has emerged as a powerful tool for efficiently sampling free energy landscapes for conformational and chemical transitions. However, daunting challenges remain in efficiently getting such simulations to scale to the very large number of replicas required to address problems in state spaces beyond two dimensions. The development of enabling technology to carry out such simulations is in its infancy, and thus it remains an open question as to which applications demand extension into higher dimensions. In the present work, we explore this problem space by applying asynchronous Hamiltonian replica exchange molecular dynamics with a combined quantum mechanical/molecular mechanical potential to explore the conformational space for a simple ribonucleoside. This is done using a newly developed software framework capable of executing >3,000 replicas with only enough resources to run 2,000 simultaneously. This may not be possible with traditional synchronous replica exchange approaches. Our results demonstrate 1.) the necessity of high dimensional sampling simulations for biological systems, even as simple as a single ribonucleoside, and 2.) the utility of asynchronous exchange protocols in managing simultaneous resource requirements expected in high dimensional sampling simulations. It is expected that more complicated systems will only increase in computational demand and complexity, and thus the reported asynchronous approach may be increasingly beneficial in order to make such applications available to a broad range of computational scientists.


international parallel and distributed processing symposium | 2015

Exploring Data Staging Across Deep Memory Hierarchies for Coupled Data Intensive Simulation Workflows

Tong Jin; Fan Zhang; Qian Sun; Hoang Bui; Melissa Romanus; Norbert Podhorszki; Scott Klasky; Hemanth Kolla; Jacqueline H. Chen; Robert Hager; Choong-Seock Chang; Manish Parashar

As applications target extreme scales, data staging and in-situ/in-transit data processing have been proposed to address the data challenges and improve scientific discovery. However, further research is necessary in order to understand how growing data sizes from data intensive simulations coupled with the limited DRAM capacity in High End Computing systems will impact the effectiveness of this approach. In this paper, we explore how we can use deep memory levels for data staging, and develop a multi-tiered data staging method that spans bothDRAM and solid state disks (SSD). This approach allows us to support both code coupling and data management for data intensive simulation workflows. We also show how an adaptive application-aware data placement mechanism can dynamically manage and optimize data placement across the DRAM ands storage levels in this multi-tiered data staging method. We present an experimental evaluation of our approach using wolf resources: an Infiniband cluster (Sith) and a Cray XK7system (Titan), and using combustion (S3D) and fusion (XGC1) simulations.


ieee international conference on high performance computing data and analytics | 2015

Adaptive data placement for staging-based coupled scientific workflows

Qian Sun; Tong Jin; Melissa Romanus; Hoang Bui; Fan Zhang; Hongfeng Yu; Hemanth Kolla; Scott Klasky; Jacqueline H. Chen; Manish Parashar

Data staging and in-situ/in-transit data processing are emerging as attractive approaches for supporting extreme scale scientific workflows. These approaches improve end-to-end performance by enabling runtime data sharing between coupled simulations and data analytics components of the workflow. However, the complex and dynamic data exchange patterns exhibited by the workflows coupled with the varied data access behaviors make efficient data placement within the staging area challenging. In this paper, we present an adaptive data placement approach to address these challenges. Our approach adapts data placement based on application-specific dynamic data access patterns, and applies access pattern-driven and location-aware mechanisms to reduce data access costs and to support efficient data sharing between the multiple workflow components. We experimentally demonstrate the effectiveness of our approach on Titan Cray XK7 using a real combustion-analyses workflow. The evaluation results demonstrate that our approach can effectively improve data access performance and overall efficiency of coupled scientific workflows.


extreme science and engineering discovery environment | 2013

A framework for flexible and scalable replica-exchange on production distributed CI

Brian K. Radak; Melissa Romanus; Emilio Gallicchio; Tai-Sung Lee; Ole Weidner; Nanjie Deng; Peng He; Wei Dai; Darrin M. York; Ronald M. Levy; Shantenu Jha

Replica exchange represents a powerful class of algorithms used for enhanced configurational and energetic sampling in a range of physical systems. Computationally it represents a type of application with multiple scales of communication. At a fine-grained level there is often communication with a replica, typically an MPI process. At a coarse-grained level, the replicas communicate with other replicas -- both temporally as well as in amount of data exchanged. This paper outlines a novel framework developed to support the flexible execution of large-scale replica exchange. The framework is flexible in the sense that it supports different coupling schemes between replicas and is agnostic to the specific underlying simulation -- classical or quantum, serial or parallel simulation. The scalability of the framework is assessed using standard simulation benchmarks. In spite of the increasing communication and coordination requirements as a function of the number of replicas, our framework supports the execution of hundreds replicas without significant overhead. Although there are several specific aspects that will benefit from further optimization, a first working prototype has the ability to fundamentally change the scale of replica exchange simulations possible on production distributed cyberinfrastructure such as XSEDE, as well as support novel usage modes. This paper also represents the release of the framework to the broader biophysical simulation community and provides details on its usage.


2016 1st Joint International Workshop on Parallel Data Storage and data Intensive Scalable Computing Systems (PDSW-DISCS) | 2016

Scientific workflows at datawarp-speed: accelerated data-intensive science using NERSC's burst buffer

Andrey Ovsyannikov; Melissa Romanus; Brian Van Straalen; Gunther H. Weber; David Trebotich

Emerging exascale systems have the ability to accelerate the time-to-discovery for scientific workflows. However, as these workflows become more complex, their generated data has grown at an unprecedented rate, making I/O constraints challenging. To address this problem advanced memory hierarchies, such as burst buffers, have been proposed as intermediate layers between the compute nodes and the parallel file system. In this paper, we utilize Cray DataWarp burst buffer coupled with in-transit processing mechanisms, to demonstrate the advantages of advanced memory hierarchies in preserving traditional coupled scientific workflows. We consider in-transit workflow which couples simulation of subsurface flows with on-the-fly flow visualization. With respect to the proposed workflow, we study the performance of the Cray DataWarp Burst Buffer and provide a comparison with the Lustre parallel file system.


extreme science and engineering discovery environment | 2012

The anatomy of successful ECSS projects: lessons of supporting high-throughput high-performance ensembles on XSEDE

Melissa Romanus; Pradeep Kumar Mantha; Matt McKenzie; Thomas C. Bishop; Emilio Gallichio; Andre Merzky; Yaakoub El Khamra; Shantenu Jha

The Extended Collaborative Support Service (ECSS) of XSEDE is a program to provide support for advanced user requirements that cannot and should not be supported via a regular ticketing system. Recently, two ECSS projects have been awarded by XSEDE management to support the high-throughput of high-performance (HTHP) molecular dynamics (MD) simulations; both of these ECSS projects use a SAGA-based Pilot-Jobs approach as the technology required to support the HTHP scenarios. Representative of the underlying ECSS philosophy, these projects were envisioned as three-way collaborations between the application stakeholders, advanced/research software development team, and the resource providers. In this paper, we describe the aims and objectives of these ECSS projects, how the deliverables have been met, and some preliminary results obtained. We believe the structure of the ECSS program enables targeted projects that address missing gaps in making distributed cyberinfrastructure systems more productive. We also describe how SAGA has been deployed on XSEDE in Community Software Area as a necessary precursor for these projects.


ieee international conference on high performance computing data and analytics | 2017

Scientific workflows at DataWarp-speed: Accelerated data-intensive science using Nersc's burst buffer

Andrey Ovsyannikov; Melissa Romanus; Brian Van Straalen; Gunther H. Weber; D Trebotich

© 2016 IEEE.Emerging exascale systems have the ability to accelerate the time-to-discovery for scientific workflows. However, as these workflows become more complex, their generated data has grown at an unprecedented rate, making I/O constraints challenging. To address this problem advanced memory hierarchies, such as burst buffers, have been proposed as intermediate layers between the compute nodes and the parallel file system. In this paper, we utilize Cray DataWarp burst buffer coupled with in-transit processing mechanisms, to demonstrate the advantages of advanced memory hierarchies in preserving traditional coupled scientific workflows. We consider in-transit workflow which couples simulation of subsurface flows with on-the-fly flow visualization. With respect to the proposed workflow, we study the performance of the Cray DataWarp Burst Buffer and provide a comparison with the Lustre parallel file system.


international workshop on data intensive distributed computing | 2016

Persistent Data Staging Services for Data Intensive In-situ Scientific Workflows

Melissa Romanus; Fan Zhang; Tong Jin; Qian Sun; Hoang Bui; Manish Parashar; Jong Youl Choi; Saloman Janhunen; R. Hager; Scott Klasky; Choong-Seock Chang; Ivan Rodero

Scientific simulation workflows executing on very large scale computing systems are essential modalities for scientific investigation. The increasing scales and resolution of these simulations provide new opportunities for accurately modeling complex natural and engineered phenomena. However, the increasing complexity necessitates managing, transporting, and processing unprecedented amounts of data, and as a result, researchers are increasingly exploring data-staging and in-situ workflows to reduce data movement and data-related overheads. However, as these workflows become more dynamic in their structures and behaviors, data staging and in-situ solutions must evolve to support new requirements. In this paper, we explore how the service-oriented concept can be applied to extreme-scale in-situ workflows. Specifically, we explore persistent data staging as a service and present the design and implementation of DataSpaces as a Service, a service-oriented data staging framework. We use a dynamically coupled fusion simulation workflow to illustrate the capabilities of this framework and evaluate its performance and scalability.


Proceedings of the Second Internationsl Workshop on Extreme Scale Programming Models and Middleware | 2016

In-staging data placement for asynchronous coupling of task-based scientific workflows

Qian Sun; Melissa Romanus; Tong Jin; Hongfeng Yu; Peer-Timo Bremer; Steve Petruzza; Scott Klasky; Manish Parashar

Coupled application workflows composed of applications implemented using task-based models present new coupling and data exchange challenges, due to the asynchronous interaction and coupling behaviors between tasks of the component applications. In this paper, we present an adaptive data placement approach that addresses these challenges by dynamically adjusting to the asynchronous coupling patterns. Specifically, it places data across a set of staging cores/nodes with an awareness of the application-specific data locality requirements and the runtime task executions at these staging cores/nodes, with the goal of reducing end-to-end execution time and data movement overhead of the workflow. We experimentally demonstrate the effectiveness of our approach on the Titan Cray XK7 system using representative data coupling patterns derived from current scientific workflows. The evaluation demonstrates that our approach efficiently improves performance by reducing the time-to-solution and increasing the quality of insights for scientific discovery.


international conference on cluster computing | 2017

TGE: Machine Learning Based Task Graph Embedding for Large-Scale Topology Mapping

Jong Youl Choi; Jeremy Logan; Matthew Wolf; George Ostrouchov; Tahsin M. Kurç; Qing Liu; Norbert Podhorszki; Scott Klasky; Melissa Romanus; Qian Sun; Manish Parashar; R.M. Churchill; Choong-Seock Chang

Task mapping is an important problem in parallel and distributed computing. The goal in task mapping is to find an optimal layout of the processes of an application (or a task) onto a given network topology. We target this problem in the context of staging applications. A staging application consists of two or more parallel applications (also referred to as staging tasks) which run concurrently and exchange data over the course of computation. Task mapping becomes a more challenging problem in staging applications, because not only data is exchanged between the staging tasks, but also the processes of a staging task may exchange data with each other. We propose a novel method, called Task Graph Embedding (TGE), that harnesses the observable graph structures of parallel applications and network topologies. TGE employs a machine learning based algorithm to find the best representation of a graph, called an embedding, onto a space in which the task-to-processor mapping problem can be solved. We evaluate and demonstrate the effectiveness of TGE experimentally with the communication patterns extracted from runs of XGC, a large-scale fusion simulation code, on Titan.

Collaboration


Dive into the Melissa Romanus's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Scott Klasky

Oak Ridge National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Andrey Ovsyannikov

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brian Van Straalen

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Choong-Seock Chang

Princeton Plasma Physics Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge