Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Milo Polte is active.

Publication


Featured researches published by Milo Polte.


ieee international conference on high performance computing data and analytics | 2009

PLFS: a checkpoint filesystem for parallel applications

John M. Bent; Garth A. Gibson; Gary Grider; Ben McClelland; Paul Nowoczynski; James Nunez; Milo Polte; Meghan Wingate

Parallel applications running across thousands of processors must protect themselves from inevitable system failures. Many applications insulate themselves from failures by checkpointing. For many applications, checkpointing into a shared single file is most convenient. With such an approach, the size of writes are often small and not aligned with file system boundaries. Unfortunately for these applications, this preferred data layout results in pathologically poor performance from the underlying file system which is optimized for large, aligned writes to non-shared files. To address this fundamental mismatch, we have developed a virtual parallel log structured file system, PLFS. PLFS remaps an applications preferred data layout into one which is optimized for the underlying file system. Through testing on PanFS, Lustre, and GPFS, we have seen that this layer of indirection and reorganization can reduce checkpoint time by an order of magnitude for several important benchmarks and real applications without any application modification.


symposium on cloud computing | 2011

YCSB++: benchmarking and performance debugging advanced features in scalable table stores

Swapnil Patil; Milo Polte; Kai Ren; Wittawat Tantisiriroj; Lin Xiao; Julio Lopez; Garth A. Gibson; Adam Fuchs; Billie Rinaldi

Inspired by Googles BigTable, a variety of scalable, semi-structured, weak-semantic table stores have been developed and optimized for different priorities such as query speed, ingest speed, availability, and interactivity. As these systems mature, performance benchmarking will advance from measuring the rate of simple workloads to understanding and debugging the performance of advanced features such as ingest speed-up techniques and function shipping filters from client to servers. This paper describes YCSB++, a set of extensions to the Yahoo! Cloud Serving Benchmark (YCSB) to improve performance understanding and debugging of these advanced features. YCSB++ includes multi-tester coordination for increased load and eventual consistency measurement, multi-phase workloads to quantify the consequences of work deferment and the benefits of anticipatory configuration optimization such as B-tree pre-splitting or bulk loading, and abstract APIs for explicit incorporation of advanced features in benchmark tests. To enhance performance debugging, we customized an existing cluster monitoring tool to gather the internal statistics of YCSB++, table stores, system services like HDFS, and operating systems, and to offer easy post-test correlation and reporting of performance behaviors. YCSB++ features are illustrated in case studies of two BigTable-like table stores, Apache HBase and Accumulo, developed to emphasize high ingest rates and finegrained security.


petascale data storage workshop | 2007

GIGA+: scalable directories for shared file systems

Swapnil Patil; Garth A. Gibson; Samuel Lang; Milo Polte

There is an increasing use of high-performance computing (HPC) clusters with thousands of compute nodes that, with the advent of multi-core CPUs, will impose a significant challenge for storage systems: The ability to scale to handle I/O generated by applications executing in parallel in tens of thousands of threads. One such challenge is building scalable directories for cluster storage - i.e., directories that can store billions to trillions of entries and handle hundreds of thousands of operations per second.


petascale data storage workshop | 2008

Comparing performance of solid state devices and mechanical disks

Milo Polte; Jiri Simsa; Garth A. Gibson

In terms of performance, solid state devices promise to be superior technology to mechanical disks. This study investigates performance of several up-to-date high-end consumer and enterprise Flash solid state devices (SSDs) and relates their performance to that of mechanical disks. For the purpose of this evaluation, the IOZone benchmark is run in single-threaded mode with varying request size and access pattern on an ext3 filesystem mounted on these devices. The price of the measured devices is then used to allow for comparison of price per performance. Measurements presented in this study offer an evaluation of cost-effectiveness of a Flash based SSD storage solution over a range of workloads. In particular, for sequential access pattern the SSDs are up to 10 times faster for reads and up to 5 times faster than the disks. For random reads, the SSDs provide up to 200times performance advantage. For random writes the SSDs provide up to 135times performance advantage. After weighting these numbers against the prices of the tested devices, we can conclude that SSDs are approaching price per performance of magnetic disks for sequential access patterns workloads and are superior technology to magnetic disks for random access patterns.


petascale data storage workshop | 2009

...and eat it too: high read performance in write-optimized HPC I/O middleware file formats

Milo Polte; Jay F. Lofstead; John M. Bent; Garth A. Gibson; Scott Klasky; Qing Liu; Manish Parashar; Norbert Podhorszki; Karsten Schwan; Meghan Wingate; Matthew Wolf

As HPC applications run on increasingly high process counts on larger and larger machines, both the frequency of checkpoints needed for fault tolerance [14] and the resolution and size of Data Analysis Dumps are expected to increase proportionally. In order to maintain an acceptable ratio of time spent performing useful computation work to time spent performing I/O, write bandwidth to the underlying storage system must increase proportionally to this increase in the checkpoint and computation size. Unfortunately, popular scientific self-describing file formats such as netCDF [8] and HDF5 [3] are designed with a focus on portability and flexibility. Extra care and careful crafting of the output structure and API calls is required to optimize for write performance using these APIs. To provide sufficient write bandwidth to continue to support the demands of scientific applications, the HPC community has developed a number of I/O middleware layers, that structure output into write-optimized file formats. However, the obvious concern with any write optimized file format would be a corresponding penalty on reads. In the log-structured filesystem [13], for example, a file generated by random writes could be written efficiently, but reading the file back sequentially later would result in very poor performance. Simulation results require efficient read-back for visualization and analytics, and though most checkpoint files are never used, the efficiency of a restart is very important in the face of inevitable failures. The utility of write speed improving middleware would be greatly diminished if it sacrificed acceptable read performance. In this paper we examine the read performance of two write-optimized middleware layers on large parallel machines and compare it to reading data natively in popular file formats.


petascale data storage workshop | 2008

Fast log-based concurrent writing of checkpoints

Milo Polte; Jiri Simsa; Wittawat Tantisiriroj; Garth A. Gibson; Shobhit Dayal; Mikhail Chainani; Dilip Kumar Uppugandla

This report describes how a file system level log-based technique can improve the write performance of many-to-one write checkpoint workload typical for high performance computations. It is shown that a simple log-based organization can provide for substantial improvements in the write performance while retaining the convenience of a single flat file abstraction. The improvement of the write performance comes at the cost of degraded read performance however. Techniques to alleviate the read performance penalty, such as file reconstruction on the first read, are discussed.


ieee international conference on high performance computing data and analytics | 2012

Poster: PLFS/HDFS: HPC Applications on Cloud Storage

Charles D. Cranor; Milo Polte; Garth A. Gibson

Long running large scale HPC applications protect themselves from failures by periodically checkpointing their state to a single file stored in a distributed network filesystem. These filesystems commonly provide a POSIX-style interface for reading and writing files. HDFS is a filesystem used in cloud computing by Apache Hadoop. HDFS is optimized for Hadoop jobs that do not require full POSIX I/O semantics. Only one process may write to an HDFS file, and all writes are appends. Our work enables multiple HPC processes to checkpoint their state into an HDFS file using PLFS. PLFS is a middleware filesystem that converts random I/O into log-based I/O. We added a new I/O store layer to PLFS that allows it to use non-POSIX filesystems like HDFS as backing store. HPC applications can now checkpoint to HDFS, allowing HPC and cloud to share the same storage systems and work with each others data.


high performance distributed computing | 2011

Six degrees of scientific data: reading patterns for extreme scale science IO

Jay F. Lofstead; Milo Polte; Garth A. Gibson; Scott Klasky; Karsten Schwan; Ron A. Oldfield; Matthew Wolf; Qing Liu


Archive | 2009

Directions for Shingled-Write and Two-Dimensional Magnetic Recording System Architectures: Synergies with Solid-State Disks

Garth A. Gibson; Milo Polte


Archive | 2009

Enabling Enterprise Solid State Disks Performance

Milo Polte; Jiri Simsa; Garth A. Gibson

Collaboration


Dive into the Milo Polte's collaboration.

Top Co-Authors

Avatar

Garth A. Gibson

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Swapnil Patil

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Jay F. Lofstead

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

Jiri Simsa

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Karsten Schwan

University of New Mexico

View shared research outputs
Top Co-Authors

Avatar

Lin Xiao

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Matthew Wolf

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Meghan Wingate

Los Alamos National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Qing Liu

University of Tennessee

View shared research outputs
Researchain Logo
Decentralizing Knowledge