Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Stephen R. Wheat is active.

Publication


Featured researches published by Stephen R. Wheat.


Archive | 1994

SUNMOS for the Intel Paragon - a brief user`s guide

Arthur B. Maccabe; R. Riesen; K.S. McCurley; Stephen R. Wheat

SUNMOS is an acronym for Sandia/UNM Operating System. It was originally developed for the nCUBE-2 MIMD supercomputer between January and December of 1991. Between April and August of 1993, SUNMOS was ported to the Intel Paragon. This document provides a quick overview of how to compile and run jobs using the SUNMOS environment on the Paragon. The primary goal of SUNMOS is to provide high performance message passing and process support an example of its capabilities, SUNMOS Release 1.4 occupies approximately 240K of memory on a Paragon node, and is able to send messages at bandwidths of 165 megabytes per second with latencies as low as 42 microseconds using Intel NX calls. By contrast, Release 1.2 of OSF/1 for the Paragon occupies approximately 7 megabytes of memory on a node, has a peak bandwidth of 65 megabytes per second, and latencies as low as 42 microseconds (the communication numbers are reported elsewhere in these proceedings).


software product lines | 1993

Out of core, out of mind: practical parallel I/O

David E. Womble; David S. Greenberg; Rolf Riesen; Stephen R. Wheat

Parallel computers are becoming more powerful and more complex in response to the demand for computing power by scientists and engineers. Inevitably, new and more complex I/O systems will be developed for these systems. In particular we believe that the I/O system must provide the programmer with the ability to explicitly manage storage (despite the trend toward complex parallel file systems and caching schemes). One method of doing so is to have a partitioned secondary storage in which each processor owns a logical disk. Along with operating system enhancements which allow overheads such as buffer copying to be avoided and libraries to support optimal remapping of data, this sort of I/O system meets the needs of high performance computing.<<ETX>>


conference on high performance computing (supercomputing) | 1993

A massively parallel adaptive finite element method with dynamic load balancing

Karen D. Devine; Joseph E. Flaherty; Stephen R. Wheat; Arthur B. Maccabe

The authors construct massively parallel adaptive finite element methods for the solution of hyperbolic conservation laws. Spatial discretization is performed by a discontinuous Galerkin finite element method using a basis of piecewise Legendre polynomials. Temporal discretization utilizes a Runge-Kutta method. Dissipative fluxes and projection limiting prevent oscillations near solution discontinuities. The resulting method is of high order and may be parallelized efficiently on MIMD computers. The authors demonstrate parallel efficiency through computations on a 1024-processor nCUBE/2 hypercube. They present results using adaptive p-refinement to reduce the computational cost of the method, and tiling, a dynamic, element-based data migration system that maintains global load balance of the adaptive method by overlapping neighborhoods of processors that each perform local balancing.


Archive | 1995

Parallel Partitioning Strategies for the Adaptive Solution of Conservation Laws

Karen D. Devine; Joseph E. Flaherty; Raymond M. Loy; Stephen R. Wheat

We describe and examine the performance of adaptive methods for solving hyperbolic systems of conservation laws on massively parallel computers. The differential system is approximated by a discontinuous Galerkin finite element method with a hierarchical Legendre piecewise polynomial basis for the spatial discretization. Fluxes at element boundaries are computed by solving an approximate Riemann problem; a projection limiter is applied to keep the average solution monotone; time discretization is performed by Runge-Kutta integration; and a p-refinement-based error estimate is used as an enrichment indicator. Adaptive order (p-) and mesh (h-) refinement algorithms are presented and demonstrated. Using an element-based dynamic load balancing algorithm called tiling and adaptive p-refinement, parallel efficiencies of over 60% are achieved on a 1024-processor nCUBE/2 hypercube. We also demonstrate a fast, tree-based parallel partitioning strategy for three-dimensional octree-structured meshes. This method produces partition quality comparable to recursive spectral bisection at a greatly reduced cost.


Scientific Programming | 1994

PUMA: an operating system for massively parallel systems

Stephen R. Wheat; Arthur B. Maccabe; Rolf Riesen; David W. Van Dresser; T. Mack Stallcup

This article presents an overview of PUMA (Performance-oriented, User-managed Messaging Architecture), a message-passing kernel for massively parallel systems. Message passing in PUMA is based on portals - an opening in the address space of an application process. Once an application process has established a portal, other processes can write values into the portal using a simple send operation. Because messages are written directly into the address space of the receiving process, there is no need to buffer messages in the PUMA kernel and later copy them into the applications address space. PUMA consists of two components: the quintessential kernel (Q-Kernel) and the process control thread (PCT). Although the PCT provides management decisions, the Q-Kernel controls access and implements the policies specified by the PCT.


conference on high performance computing (supercomputing) | 1994

Applications of boundary element methods on the Intel Paragon

David E. Womble; David S. Greenberg; Stephen R. Wheat; Robert E. Benner; Marc S. Ingber; Greg Henry; Satya Gutpa

This paper describes three applications of the boundary element method and their implementations on the Intel Paragon supercomputer. Each of these applications sustains over 99 Gflops/s based on wall-clock time for the entire application and an actual count of flops executed; one application sustains over 140 Gflops/s. Each application accepts the description of an arbitrary geometry and computes the solution to a problem of commercial and research interest. The common kernel for these applications is a dense equation solver based on LU factorization. It is generally accepted that good performance can be achieved by dense matrix algorithms, but achieving the excellent performance demonstrated here required the development of a variety of special techniques to take full advantage of the power of the Intel Paragon.<<ETX>>


conference on high performance computing (supercomputing) | 2006

Addressing high performance and grid challenges: Intel and CERN

Stephen R. Wheat; Bob Jones

Intel will briefly introduce is new HPC products and upgrades, and future HPC plans and roadmaps and challenges while inviting director of the CERN EGEE Project to demonstrate successful relationship between Intel and this leading scientific organization.


conference on high performance computing (supercomputing) | 2006

What's inside the grid? a discussion of standards and the future of computing

Gary Tyreman; Mark Linesch; Stephen R. Wheat; Andre Hill

Cluster computing is a disruptive force that has quickly reshaped the HPC market and rapidly gained acceptance as a solution to an ever-increasing number of business and research problems in the data center. To sustain, and accelerate the growth of grid and related fields, we need a set of reference architectures, implementations and technology standards that will enable interoperable components and an ubiquitous computing experience.This panel will debut the initiative to establish standards to discuss and implement grid systems. Industry and academic representatives will debate current challenges in grid computing, why we need standards, how we should talk about grid, what grid will mean to the industry for the next ten years, and how grid is a spring board for academics to take HPC, virtualization, and distributed computing to new heights. During the Q&A period, attendees will have an opportunity to voice their concerns or support of initiatives addressed.


Archive | 1993

Dynamic load balancing of applications

Stephen R. Wheat


Archive | 1993

Beyond core: Making parallel computer I/O practical

D. Womble; D. Greenberg; Stephen R. Wheat; R. Riesen

Collaboration


Dive into the Stephen R. Wheat's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

David E. Womble

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

David S. Greenberg

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

Joseph E. Flaherty

Rensselaer Polytechnic Institute

View shared research outputs
Top Co-Authors

Avatar

Karen D. Devine

Rensselaer Polytechnic Institute

View shared research outputs
Top Co-Authors

Avatar

Marc S. Ingber

University of New Mexico

View shared research outputs
Top Co-Authors

Avatar

Raymond M. Loy

Rensselaer Polytechnic Institute

View shared research outputs
Researchain Logo
Decentralizing Knowledge