Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Amanda Peters is active.

Publication


Featured researches published by Amanda Peters.


computing frontiers | 2007

Parallel genomic sequence-search on a massively parallel system

Oystein Thorsen; Brian E. Smith; Carlos P. Sosa; Karl Jiang; Heshan Lin; Amanda Peters; Wu-chun Feng

In the life sciences, genomic databases for sequence search have been growing exponentially in size. As a result, faster sequence-search algorithms to search these databases continue to evolve to cope with algorithmic time complexity. The ubiquitous tool for such search is the Basic Local Alignment Search Tool (BLAST) [1] from the National Center for Biotechnology Information (NCBI). Despite continued algorithmic improvements in BLAST, it cannot keep up with the rate at which the database is exponentially increasing in size. Therefore, parallel implement-ations such as mpiBLAST have emerged to address this problem. The performance of such implementations depends on a myriad of factors including algorithmic, architectural, and mapping of the algorithm to the architecture. This paper describes modifications and extensions to a parallel and distributed-memory version of BLAST called mpiBLAST-PIO and how it maps to a massively parallel system, specifically IBM Blue Gene/L (BG/L). The extensions include a virtual file manager, a multiple master run-time model, efficient fragment distribution, and intelligent load balancing. In this study, we have shown that our optimized mpiBLAST-PIO on BG/L using a query with 28014 sequences and the NR and NT databases scales to 8192 nodes (two cores per node). The cases tested here are well suited for a massively parallel system.


IEEE Transactions on Parallel and Distributed Systems | 2008

An Efficient Parallel Implementation of the Hidden Markov Methods for Genomic Sequence-Search on a Massively Parallel System

Karl Jiang; Oystein Thorsen; Amanda Peters; Brian E. Smith; Carlos P. Sosa

Bioinformatics databases used for sequence comparison and sequence alignment are growing exponentially. This has popularized programs that carry out database searches. Current implementations of sequence alignment methods based on hidden Markov models (HMM) have proven to be computationally intensive and, hence, amenable to architectures with multiple processors. In this paper, we describe a modified version of the original parallel implementation of HMMs on a massively parallel system. This is part of the HMMER bioinformatics code. HMMER 2.3.2 uses profile HMMs for sensitive database searching based on statistical descriptions of a sequence familys consensus (Durbin et al., 1998), Two of the nine programs were further parallelized to take advantage of the large number of processors, namely, hmmsearch and hmmpfam. For our study, we start by porting the parallel virtual machine (PVM) versions of these two programs currently available as part of the HMMER suite of programs. We report the performance of these nonoptimized versions as baselines. Our work also includes the introduction of an alternate sequence file indexing, multiple-master configuration, dynamic data collection and, finally, load balancing via the indexed sequence files. This set of optimizations constitutes our modified version for massively parallel systems. Our results show parallel performance improvements of more than one order of magnitude (16 times) for hmmsearch and hmmpfam.


international parallel and distributed processing symposium | 2008

Asynchronous task dispatch for high throughput computing for the eServer IBM Blue Gene® Supercomputer

Amanda Peters; Alan King; Tom Budnik; Patrick McCarthy; Paul Michaud; Mike Mundy; James C. Sexton; Greg Stewart

High Throughput Computing (HTC) environments strive to provide large amounts of processing capacity to customers over long periods of time by exploiting existing resources on the network according to Basney and Livny [1]. A single Blue Gene/L rack can provide thousands of CPU resources into HTC environments. This paper discusses the implementation of an asynchronous task dispatch system that exploits a recently released feature of the Blue Gene/L control system - called HTC mode - and presents data on experimental runs consisting of the asynchronous submission of multiple batches of thousands of tasks for financial workloads. The methodology developed here demonstrates how systems with very large processor counts and light-weight kernels can be configured to deliver capacity computing at the individual processor level in future petascale computing systems.


Ibm Journal of Research and Development | 2008

EUDOC on the IBM Blue Gene/L system: accelerating the transfer of drug discoveries from laboratory to patient

Yuan-Ping Pang; Timothy J. Mullins; Brent Allen Swartz; Jeff S. McAllister; Brian E. Smith; Charles J. Archer; Roy Glenn Musselman; Amanda Peters; Brian Paul Wallenfelt; Kurt Walter Pinnow

EUDOC™ is a molecular docking program that has successfully helped to identify new drug leads. This virtual screening (VS) tool identifies drug candidates by computationally testing the binding of these drugs to biologically important protein targets. This approach can reduce the research time required of biochemists, accelerating the identification of therapeutically useful drugs and helping to transfer discoveries from the laboratory to the patient. Migration of the EUDOC application code to the IBM Blue Gene/L™ (BG/L) supercomputer has been highly successful. This migration led to a 200-fold improvement in elapsed time for a representative VS application benchmark. Three focus areas provided benefits. First, we enhanced the performance of serial code through application redesign, hand-tuning, and increased usage of SIMD (single-instruction, multiple-data) floating-point unit operations. Second, we studied computational load-balancing schemes to maximize processor utilization and application scalability for the massively parallel architecture of the BG/L system. Third, we greatly enhanced system I/O interaction design. We also identified and resolved severe performance bottlenecks, allowing for efficient performance on more than 4,000 processors. This paper describes specific improvements in each of the areas of focus.


Archive | 2007

Dynamically rerouting node traffic on a massively parallel computer system using hint bits

David L. Darrington; Patrick McCarthy; Amanda Peters; Albert Sidelnik; Brian E. Smith; Brent Allen Swartz


Archive | 2008

High Throughput Computing Validation for Drug Discovery Using the DOCK Program on a Massively Parallel System

Amanda Peters; Marcus E. Lundberg; P. Therese Lang; Carlos P. Sosa


Archive | 2007

SHARING LOADED JAVA CLASSES AMONG A PLURALITY OF NODES

Eric Lawrence Barsness; David L. Darrington; Amanda Peters; John Matthew Santosuosso


Archive | 2007

Method and Apparatus for Operating a Massively Parallel Computer System to Utilize Idle Processor Capability at Process Synchronization Points

Charles J. Archer; Roy Glenn Musselman; Amanda Peters; Kurt Walter Pinnow; Brent Allen Swartz


Archive | 2007

Optimization of job distribution on a multi-node computer system

Eric Lawrence Barness; David L. Darrington; Amanda Peters; John Matthew Santosuosso


Archive | 2008

Process Migration Based on Exception Handling in a Multi-Node Environment

Eric Lawrence Barsness; David L. Darrington; Amanda Peters; John Matthew Santosuosso

Collaboration


Dive into the Amanda Peters's collaboration.

Researchain Logo
Decentralizing Knowledge