Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Philip J. Maechling is active.

Publication


Featured researches published by Philip J. Maechling.


international conference on e-science | 2009

Scientific workflow applications on Amazon EC2

Gideon Juve; Ewa Deelman; Karan Vahi; Gaurang Mehta; G. Bruce Berriman; Benjamin P. Berman; Philip J. Maechling

The proliferation of commercial cloud computing providers has generated significant interest in the scientific computing community. Much recent research has attempted to determine the benefits and drawbacks of cloud computing for scientific applications. Although clouds have many attractive features, such as virtualization, on-demand provisioning, and “pay as you go” usage-based pricing, it is not clear whether they are able to deliver the performance required for scientific applications at a reasonable price. In this paper we examine the performance and cost of clouds from the perspective of scientific workflow applications. We use three characteristic workflows to compare the performance of a commercial cloud with that of a typical HPC system, and we analyze the various costs associated with running those workflows in the cloud. We find that the performance of clouds is not unreasonable given the hardware resources provided, and that performance comparable to HPC systems can be achieved given similar resources. We also find that the cost of running workflows on a commercial cloud can be reduced by storing data in the cloud rather than transferring it from outside.


Future Generation Computer Systems | 2015

Pegasus, a workflow management system for science automation

Ewa Deelman; Karan Vahi; Gideon Juve; Mats Rynge; Scott Callaghan; Philip J. Maechling; Rajiv Mayani; Weiwei Chen; Rafael Ferreira da Silva; Miron Livny; Kent Wenger

Modern science often requires the execution of large-scale, multi-stage simulation and data analysis pipelines to enable the study of complex systems. The amount of computation and data involved in these pipelines requires scalable workflow management systems that are able to reliably and efficiently coordinate and automate data movement and task execution on distributed computational resources: campus clusters, national cyberinfrastructures, and commercial and academic clouds. This paper describes the design, development and evolution of the Pegasus Workflow Management System, which maps abstract workflow descriptions onto distributed computing infrastructures. Pegasus has been used for more than twelve years by scientists in a wide variety of domains, including astronomy, seismology, bioinformatics, physics and others. This paper provides an integrated view of the Pegasus system, showing its capabilities that have been developed over time in response to application needs and to the evolution of the scientific computing platforms. The paper describes how Pegasus achieves reliable, scalable workflow execution across a wide variety of computing infrastructures. Comprehensive description of the Pegasus Workflow Management System.Detailed explanation of Pegasus workflow transformations.Data management in Pegasus.Earthquake science application example.


international conference on e science | 2006

Managing Large-Scale Workflow Execution from Resource Provisioning to Provenance Tracking: The CyberShake Example

Ewa Deelman; Scott Callaghan; Edward H. Field; H. Francoeur; Robert W. Graves; Nitin Gupta; Vipin Gupta; Thomas H. Jordan; Carl Kesselman; Philip J. Maechling; John Mehringer; Gaurang Mehta; David A. Okaya; Karan Vahi; Li Zhao

This paper discusses the process of building an environment where large-scale, complex, scientific analysis can be scheduled onto a heterogeneous collection of computational and storage resources. The example application is the Southern California Earthquake Center (SCEC) CyberShake project, an analysis designed to compute probabilistic seismic hazard curves for sites in the Los Angeles area. We explain which software tools were used to build to the system, describe their functionality and interactions. We show the results of running the CyberShake analysis that included over 250,000 jobs using resources available through SCEC and the TeraGrid.


ieee international conference on high performance computing data and analytics | 2010

Data Sharing Options for Scientific Workflows on Amazon EC2

Gideon Juve; Ewa Deelman; Karan Vahi; Gaurang Mehta; G. Bruce Berriman; Benjamin P. Berman; Philip J. Maechling

Efficient data management is a key component in achieving good performance for scientific workflows in distributed environments. Workflow applications typically communicate data between tasks using files. When tasks are distributed, these files are either transferred from one computational node to another, or accessed through a shared storage system. In grids and clusters, workflow data is often stored on network and parallel file systems. In this paper we investigate some of the ways in which data can be managed for workflows in the cloud. We ran experiments using three typical workflow applications on Amazons EC2. We discuss the various storage and file systems we used, describe the issues and problems we encountered deploying them on EC2, and analyze the resulting performance and cost of the workflows.


ieee international conference on high performance computing data and analytics | 2010

Scalable Earthquake Simulation on Petascale Supercomputers

Yifeng Cui; Kim B. Olsen; Thomas H. Jordan; Kwangyoon Lee; Jun Zhou; Patrick Small; D. Roten; Geoffrey Palarz Ely; Dhabaleswar K. Panda; Amit Chourasia; John M. Levesque; Steven M. Day; Philip J. Maechling

Petascale simulations are needed to understand the rupture and wave dynamics of the largest earthquakes at shaking frequencies required to engineer safe structures (> 1 Hz). Toward this goal, we have developed a highly scalable, parallel application (AWP-ODC) that has achieved “M8”: a full dynamical simulation of a magnitude-8 earthquake on the southern San Andreas fault up to 2 Hz. M8 was calculated using a uniform mesh of 436 billion 40-m3 cubes to represent the three-dimensional crustal structure of Southern California, in a 800 km by 400 km area, home to over 20 million people. This production run producing 360 sec of wave propagation sustained 220 Tflop/s for 24 hours on NCCS Jaguar using 223,074 cores. As the largest-ever earthquake simulation, M8 opens new territory for earthquake science and engineering—the physics-based modeling of the largest seismic hazards with the goal of reducing their potential for loss of life and property.


Bulletin of the Seismological Society of America | 2008

TeraShake2: Spontaneous Rupture Simulations of Mw 7.7 Earthquakes on the Southern San Andreas Fault

Kim B. Olsen; Steven M. Day; Jean-Bernard Minster; Yifeng Cui; Amit Chourasia; David A. Okaya; Philip J. Maechling; Thomas H. Jordan

Abstract Previous numerical simulations (TeraShake1) of large ( M w 7.7) southern San Andreas fault earthquakes predicted localized areas of strong amplification in the Los Angeles area associated with directivity and wave-guide effects from northwestward-propagating rupture scenarios. The TeraShake1 source was derived from inversions of the 2002 M w 7.9 Denali, Alaska, earthquake. That source was relatively smooth in its slip distribution and rupture characteristics, owing both to resolution limits of the inversions and simplifications imposed by the kinematic parameterization. New simulations (TeraShake2), with a more complex source derived from spontaneous rupture modeling with small-scale stress-drop heterogeneity, predict a similar spatial pattern of peak ground velocity (PGV), but with the PGV extremes decreased by factors of 2–3 relative to TeraShake1. The TeraShake2 source excites a less coherent wave field, with reduced along-strike directivity accompanied by streaks of elevated ground motion extending away from the fault trace. The source complexity entails abrupt changes in the direction and speed of rupture correlated to changes in slip-velocity amplitude and waveform, features that might prove challenging to capture in a purely kinematic parameterization. Despite the reduced PGV extremes, northwest-rupturing TeraShake2 simulations still predict entrainment by basin structure of a strong directivity pulse, with PGVs in Los Angeles and San Gabriel basins that are much higher than predicted by empirical methods. Significant areas of those basins have predicted PGV above the 2% probability of exceedance (POE) level relative to current attenuation relationships (even when the latter includes a site term to account for local sediment depth), and wave-guide focusing produces localized areas with PGV at roughly 0.1%–0.2% POE (about a factor of 4.5 above the median). In contrast, at rock sites in the 0–100-km distance range, the median TeraShake2 PGVs are in very close agreement with the median empirical prediction, and extremes nowhere reach the 2% POE level. The rock-site agreement lends credibility to some of our source-modeling assumptions, including overall stress-drop level and the manner in which we assigned dynamic parameters to represent the mechanical weakness of near-surface material. Future efforts should focus on validating and refining these findings, assessing their probabilities of occurrence relative to alternative rupture scenarios for the southern San Andreas fault, and incorporating them into seismic hazard estimation for southern California.


grid computing | 2012

An Evaluation of the Cost and Performance of Scientific Workflows on Amazon EC2

Gideon Juve; Ewa Deelman; G. Bruce Berriman; Benjamin P. Berman; Philip J. Maechling

Workflows are used to orchestrate data-intensive applications in many different scientific domains. Workflow applications typically communicate data between processing steps using intermediate files. When tasks are distributed, these files are either transferred from one computational node to another, or accessed through a shared storage system. As a result, the efficient management of data is a key factor in achieving good performance for workflow applications in distributed environments. In this paper we investigate some of the ways in which data can be managed for workflows in the cloud. We ran experiments using three typical workflow applications on Amazon’s EC2 cloud computing platform. We discuss the various storage and file systems we used, describe the issues and problems we encountered deploying them on EC2, and analyze the resulting performance and cost of the workflows.


international conference on management of data | 2005

Simplifying construction of complex workflows for non-expert users of the Southern California Earthquake Center Community Modeling Environment

Philip J. Maechling; Hans Chalupsky; Maureen Dougherty; Ewa Deelman; Yolanda Gil; Sridhar Gullapalli; Vipin Gupta; Carl Kesselman; Jihie Kim; Gaurang Mehta; Brian Mendenhall; Thomas A. Russ; Gurmeet Singh; Marc Spraragen; Garrick Staples; Karan Vahi

Workflow systems often present the user with rich interfaces that express all the capabilities and complexities of the application programs and the computing environments that they support. However, non-expert users are better served with simple interfaces that abstract away system complexities and still enable them to construct and execute complex workflows. To explore this idea, we have created a set of tools and interfaces that simplify the construction of workflows. Implemented as part of the Community Modeling Environment developed by the Southern California Earthquake Center, these tools, are integrated into a comprehensive workflow system that supports both domain experts as well as non expert users.


Archive | 2007

SCEC CyberShake Workflows—Automating Probabilistic Seismic Hazard Analysis Calculations

Philip J. Maechling; Ewa Deelman; Li Zhao; Robert W. Graves; Gaurang Mehta; Nitin Gupta; John Mehringer; Carl Kesselman; Scott Callaghan; David A. Okaya; H. Francoeur; Vipin Gupta; Yifeng Cui; Karan Vahi; Thomas H. Jordan; Edward H. Field

The Southern California Earthquake Center (SCEC) is a community of more than 400 scientists from over 54 research organizations that conducts geophysical research in order to develop a physics-based understanding of earthquake processes and to reduce the hazard from earthquakes in the Southern California region [377].


Archive | 2014

CISN ShakeAlert: An Earthquake Early Warning Demonstration System for California

Maren Böse; Richard M. Allen; H. Brown; G. Gua; M. Fischer; Egill Hauksson; T. Heaten; Margaret Hellweg; M. Liukis; D. S. Neuhauser; Philip J. Maechling; K. Solanki; M. Vinci; Ivan Henson; O. N. Khainovski; S. Kuyuk; M. Carpio; M.-A. Meier; Thomas H. Jordan

To demonstrate the feasibility of earthquake early warning (EEW) in California, we have developed and implemented the CISN ShakeAlert demonstration system. A Decision Module combines estimates and uncertainties determined by three algorithms implemented in parallel, \(\tau _\mathrm{{c}}-\mathrm{{P}}_\mathrm{{d}}\) Onsite, Virtual Seismologist, and ElarmS, to calculate and report at a given time the most probable earthquake magnitude and location, as well as the likelihood of correct alarm. A User Display receives the alert messages in real-time, calculates the expected local shaking intensity, and displays the information on a map. Currently, CISN ShakeAlert is being tested by \(\sim \)70 individuals and test users from industries and emergency response organizations in California. During the next 3 years we plan to expand this demonstration warning system to the entire US West Coast.

Collaboration


Dive into the Philip J. Maechling's collaboration.

Top Co-Authors

Avatar

Thomas H. Jordan

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Kim B. Olsen

San Diego State University

View shared research outputs
Top Co-Authors

Avatar

Ewa Deelman

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Robert W. Graves

United States Geological Survey

View shared research outputs
Top Co-Authors

Avatar

David A. Okaya

United States Geological Survey

View shared research outputs
Top Co-Authors

Avatar

Gaurang Mehta

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Carl Kesselman

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Gideon Juve

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Scott Callaghan

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Yifeng Cui

University of California

View shared research outputs
Researchain Logo
Decentralizing Knowledge