Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where nan Prabhat is active.

Publication


Featured researches published by nan Prabhat.


Lawrence Berkeley National Laboratory | 2009

FastBit: interactively searching massive data

Kesheng Wu; Sean Ahern; Edward W Bethel; Jacqueline H. Chen; Hank Childs; E. Cormier-Michel; Cameron Geddes; Junmin Gu; Hans Hagen; Bernd Hamann; Wendy S. Koegler; Jerome Lauret; Jeremy S. Meredith; Peter Messmer; Ekow J. Otoo; V Perevoztchikov; A. M. Poskanzer; Prabhat; Oliver Rübel; Arie Shoshani; Alexander Sim; Kurt Stockinger; Gunther H. Weber; W. M. Zhang

As scientific instruments and computer simulations produce more and more data, the task of locating the essential information to gain insight becomes increasingly difficult. FastBit is an efficient software tool to address this challenge. In this article, we present a summary of the key underlying technologies, namely bitmap compression, encoding, and binning. Together these techniques enable FastBit to answer structured (SQL) queries orders of magnitude faster than popular database systems. To illustrate how FastBit is used in applications, we present three examples involving a high-energy physics experiment, a combustion simulation, and an accelerator simulation. In each case, FastBit significantly reduces the response time and enables interactive exploration on terabytes of data.


IEEE Computer Graphics and Applications | 2010

Extreme Scaling of Production Visualization Software on Diverse Architectures

Hank Childs; David Pugmire; Sean Ahern; Brad Whitlock; Mark Howison; Prabhat; Gunther H. Weber; E. Wes Bethel

This article presents the results of experiments studying how the pure-parallelism paradigm scales to massive data sets, including 16,000 or more cores on trillion-cell meshes, the largest data sets published to date in the visualization literature. The findings on scaling characteristics and bottlenecks contribute to understanding how pure parallelism will perform in the future.


ieee international conference on high performance computing data and analytics | 2011

Parallel index and query for large scale data analysis

Jerry Chi-Yuan Chou; Mark Howison; Brian Austin; Kesheng Wu; Ji Qiang; E. Wes Bethel; Arie Shoshani; Oliver Rübel; Prabhat; Robert D. Ryne

Modern scientific datasets present numerous data management and analysis challenges. State-of-the-art index and query technologies are critical for facilitating interactive exploration of large datasets, but numerous challenges remain in terms of designing a system for processing general scientific datasets. The system needs to be able to run on distributed multi-core platforms, efficiently utilize underlying I/O infrastructure, and scale to massive datasets. We present FastQuery, a novel software framework that address these challenges. FastQuery utilizes a state-of-the- art index and query technology (FastBit) and is designed to process massive datasets on modern supercomputing plat- forms. We apply FastQuery to processing of a massive 50TB dataset generated by a large scale accelerator modeling code. We demonstrate the scalability of the tool to 11,520 cores. Motivated by the scientific need to search for interesting particles in this dataset, we use our framework to reduce search time from hours to tens of seconds.


Journal of Climate | 2015

Resolution Dependence of Future Tropical Cyclone Projections of CAM5.1 in the U.S. CLIVAR Hurricane Working Group Idealized Configurations

Michael F. Wehner; Prabhat; Kevin A. Reed; Dáithí A. Stone; William D. Collins; Julio T. Bacmeister

AbstractThe four idealized configurations of the U.S. CLIVAR Hurricane Working Group are integrated using the global Community Atmospheric Model version 5.1 at two different horizontal resolutions, approximately 100 and 25 km. The publicly released 0.9° × 1.3° configuration is a poor predictor of the sign of the 0.23° × 0.31° model configuration’s change in the total number of tropical storms in a warmer climate. However, it does predict the sign of the higher-resolution configuration’s change in the number of intense tropical cyclones in a warmer climate. In the 0.23° × 0.31° model configuration, both increased CO2 concentrations and elevated sea surface temperature (SST) independently lower the number of weak tropical storms and shorten their average duration. Conversely, increased SST causes more intense tropical cyclones and lengthens their average duration, resulting in a greater number of intense tropical cyclone days globally. Increased SST also increased maximum tropical storm instantaneous precip...


ieee international conference on high performance computing data and analytics | 2013

Taming parallel I/O complexity with auto-tuning

Babak Behzad; Huong Luu; Joseph Huchette; Surendra Byna; Prabhat; Ruth A. Aydt; Quincey Koziol; Marc Snir

We present an auto-tuning system for optimizing I/O performance of HDF5 applications and demonstrate its value across platforms, applications, and at scale. The system uses a genetic algorithm to search a large space of tunable parameters and to identify effective settings at all layers of the parallel I/O stack. The parameter settings are applied transparently by the auto-tuning system via dynamically intercepted HDF5 calls. To validate our auto-tuning system, we applied it to three I/O benchmarks (VPIC, VORPAL, and GCRM) that replicate the I/O activity of their respective applications. We tested the system with different weak-scaling configurations (128, 2048, and 4096 CPU cores) that generate 30 GB to 1 TB of data, and executed these configurations on diverse HPC platforms (Cray XE6, IBM BG/P, and Dell Cluster). In all cases, the auto-tuning framework identified tunable parameters that substantially improved write performance over default system settings. We consistently demonstrate I/O write speedups between 2× and 100× for test configurations.


ieee international conference on high performance computing data and analytics | 2008

High performance multivariate visual data exploration for extremely large data

Oliver Rübel; Prabhat; Kesheng Wu; Hank Childs; Jeremy S. Meredith; Cameron Geddes; E. Cormier-Michel; Sean Ahern; Gunther H. Weber; Peter Messmer; Hans Hagen; Bernd Hamann; E. Wes Bethel

One of the central challenges in modern science is the need to quickly derive knowledge and understanding from large, complex collections of data. We present a new approach that deals with this challenge by combining and extending techniques from high performance visual data analysis and scientific data management. This approach is demonstrated within the context of gaining insight from complex, time-varying datasets produced by a laser wakefield accelerator simulation. Our approach leverages histogram-based parallel coordinates for both visual information display as well as a vehicle for guiding a data mining operation. Data extraction and subsetting are implemented with state-of-the-art index/query technology. This approach, while applied here to accelerator science, is generally applicable to a broad set of science applications, and is implemented in a production-quality visual data analysis infrastructure. We conduct a detailed performance analysis and demonstrate good scalability on a distributed memory Cray XT4 system.


ieee international conference on high performance computing data and analytics | 2012

Parallel I/O, analysis, and visualization of a trillion particle simulation

Surendra Byna; J. Chou; Oliver Rübel; Prabhat; H. Karimabadi; W. S. Daughter; V. Roytershteyn; E. W. Bethel; Mark Howison; Ke-Jou Hsu; Kuan-Wu Lin; Arie Shoshani; A. Uselton; Kesheng Wu

Petascale plasma physics simulations have recently entered the regime of simulating trillions of particles. These unprecedented simulations generate massive amounts of data, posing significant challenges in storage, analysis, and visualization. In this paper, we present parallel I/O, analysis, and visualization results from a VPIC trillion particle simulation running on 120,000 cores, which produces ~30TB of data for a single timestep. We demonstrate the successful application of H5Part, a particle data extension of parallel HDF5, for writing the dataset at a significant fraction of system peak I/O rates. To enable efficient analysis, we develop hybrid parallel FastQuery to index and query data using multi-core CPUs on distributed memory hardware. We show good scalability results for the FastQuery implementation using up to 10,000 cores. Finally, we apply this indexing/query-driven approach to facilitate the first-ever analysis and visualization of the trillion particle dataset.


high performance distributed computing | 2014

Improving parallel I/O autotuning with performance modeling

Babak Behzad; Surendra Byna; Stefan M. Wild; Prabhat; Marc Snir

Various layers of the parallel I/O subsystem offer tunable parameters for improving I/O performance on large-scale computers. However, searching through a large parameter space is challenging. We are working towards an autotuning framework for determining the parallel I/O parameters that can achieve good I/O performance for different data write patterns. In this paper, we characterize parallel I/O and discuss the development of predictive models for use in effectively reducing the parameter space. Applying our technique on tuning an I/O kernel derived from a large-scale simulation code shows that the search time can be reduced from 12 hours to 2 hours, while achieving 54X I/O performance speedup.


Journal of Statistical Software | 2015

Parallelizing Gaussian Process Calculations in R

Christopher J. Paciorek; Benjamin Lipshitz; Wei Zhuo; Prabhat; Cari G. Kaufman; Rollin C. Thomas

We consider parallel computation for Gaussian process calculations to overcome computational and memory constraints on the size of datasets that can be analyzed. Using a hybrid parallelization approach that uses both threading (shared memory) and message-passing (distributed memory), we implement the core linear algebra operations used in spatial statistics and Gaussian process regression in an R package called bigGP that relies on C and MPI. The approach divides the covariance matrix into blocks such that the computational load is balanced across processes while communication between processes is limited. The package provides an API enabling R programmers to implement Gaussian process-based methods by using the distributed linear algebra operations without any C or MPI coding. We illustrate the approach and software by analyzing an astrophysics dataset with n = 67, 275 observations.


international workshop on data intensive distributed computing | 2012

Experiences with 100Gbps network applications

Mehmet Balman; Eric Pouyoul; Yushu Yao; E. Wes Bethel; Burlen Loring; Prabhat; John Shalf; Alex Sim; Brian Tierney

100Gbps networking has finally arrived, and many research and educational institutions have begun to deploy 100Gbps routers and services. ESnet and Internet2 worked together to make 100Gbps networks available to researchers at the Supercomputing 2011 conference in Seattle Washington. In this paper, we describe two of the first applications to take advantage of this network. We demonstrate a visualization application that enables remotely located scientists to gain insights from large datasets. We also demonstrate climate data movement and analysis over the 100Gbps network. We describe a number of application design issues and host tuning strategies necessary for enabling applications to scale to 100Gbps rates.

Collaboration


Dive into the nan Prabhat's collaboration.

Top Co-Authors

Avatar

E. Wes Bethel

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Oliver Rübel

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Surendra Byna

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Michael F. Wehner

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Gunther H. Weber

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kesheng Wu

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Cameron Geddes

Lawrence Berkeley National Laboratory

View shared research outputs
Top Co-Authors

Avatar

E. Cormier-Michel

Lawrence Berkeley National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge