Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Susan Hinrichs is active.

Publication


Featured researches published by Susan Hinrichs.


acm symposium on parallel algorithms and architectures | 1994

An architecture for optimal all-to-all personalized communication

Susan Hinrichs; Corey Kosak; David R. O'Hallaron; Thomas M. Stricker; Riichiro Take

In all-to-all personalized communication (AAPC), every node of a parallel system sends a potentially unique packet to every other node. AAPC is an important primitive operation for modern parallel compilers, since it is used to redistribute data structures during parallel computations. As an extremely dense communication pattern, AAPC causes congestion in many types of networks and therefore executes very poorly on general purpose, asynchronous message passsing routers. We present and evaluate a network architecture that executes all-to-all communication optimally on a two-dimensional torus. The router combines optimal partitions of the AAPC step with a self-synchronizing switching mechanism integrated into a conventional wormhole router. Optimality is achieved by routing along shortest paths while fully utilizing all links. A simple hardware addition for synchronized message switching can guarantee optimal AAPC routing in many existing network architectures. The flexible communication agent of the iWarp VLSI component allowed us to implement an efficient prototype for the evaluation of the hardware complexity as well as possible software overheads. The measured performance on an 8 × 8 torus exceeded 2 GigaBytes/sec or 80% of the limit set by the raw speed of the interconnects. We make a quantitative comparison of the AAPC router with a conventional message passing system. The potential gain of such a router for larger parallel programs is illustrated with the example of a two-dimensional Fast Fourier Transform.


international conference on supercomputing | 1995

Decoupling synchronization and data transfer in message passing systems of parallel computers

Thomas M. Stricker; James M. Stichnoth; David R. O'Hallaron; Susan Hinrichs; Thomas R. Gross

Synchronization is an important issue for the design of a scalable parallel computer, and some systems include special hardware support for control messages or barriers. The cost of synchronization has a high impact on the design of the message passing (communication) services. In this paper, we investigate three different communication libraries that are tailored toward the synchronization services available: (1) a version of generic send-receive message passing (PVM), which relies on traditional flow control and buffering to synchronize the data transfers; (2) message passing with pulling, i.e. a message is transferred only when the recipient is ready and requests it (as, e.g., used in NX for large messages); and (3) the decoupled direct deposit message passing that uses separate, global synchronization to ensure that nodes send messages only when the message data can be deposited directly into the final destination in the memory of the remote recipient. Measurements of these three styles on a Cray T3D demonstrate the benefits of the decoupled message passing with direct deposit. The performance advantage of this style is made possible by (1) preemptive synchronization to avoid unnecessary copies of the data, (2) high-speed barrier synchronization, and (3) improved congestion control in the network. The designers of the communication system of future parallel computers are therefore strongly encouraged to provide good synchronizationfacilities in addition to high throughput data transfers to support high performance message passing.


IEEE Parallel & Distributed Technology: Systems & Applications | 1995

Simplifying connection-based communication

Susan Hinrichs

Connection-based communication can improve performance of programs on distributed-memory systems, but it complicates programming. By defining connections at compile time, the Programmed Communication Service tool chain simplifies creation of these programs. >


IEEE Computer | 1994

Communication styles for parallel systems

Thomas R. Gross; Susan Hinrichs; David R. O'Hallaron; Thomas M. Stricker; Atsushi Hasegawa

Distributed-memory parallel systems rely on explicit message exchange for communication, but the communication operations they support can differ in many aspects. One key difference is the way messages are generated or consumed. With systolic communication, a message is transmitted as it is generated. For example, the result computed by the multiplier is sent directly to the communication subsystem for transmission to another node. With memory communication, the complete message is generated and stored in memory, and then transmitted to its destination. Since sender and receiver nodes are individually controlled, they can use different communication styles. One example of memory communication is message passing: both the sender and receiver buffer the message in memory. These two communication styles place different demands on processor design. This article illustrates each styles effect on processor resources for some key application kernels. We are targeting the iWarp system because it supports both communication styles. Two parallel-program generators, one for each communication style, automatically map the sample programs.<<ETX>>


languages and compilers for parallel computing | 1992

Utilizing New Communication Features in Compiliation for Private-Memory Machines

Susan Hinrichs; Thomas R. Gross

The communication system of some 3rd generation private-memory machines provides long-lived connections (which reserve communication resources like buffers between nodes) as well as direct access by the computation unit(s) of the node to the communication system. These features allow a compiler to find innovative solutions when compiling data-parallel programs for a private-memory machine. In this paper, we discuss some optimizations that can be included in a compiler for the iWarp system, an example private-memory parallel system with a novel communication architecture.


Archive | 1994

A Compiler Time Model for Composing Parallel Programs

Susan Hinrichs


Archive | 1995

Compiler directed architecture-dependent communication optimizations

Susan Hinrichs; Thomas R. Gross


Archive | 1992

Debugging a Parallel Program: Capturing Inter-Processor Communication in an iWarp Torus

Thomas R. Gross; Susan Hinrichs


Concurrency and Computation: Practice and Experience | 1997

Connection resource management for compiler-generated communication

Susan Hinrichs


SIGPLAN Workshop | 1992

Compiling Task and Data Parallel Programs for iWrap (Extended Abstract).

Thomas R. Gross; Susan Hinrichs; David R. O'Hallaron; James M. Stichnoth; Jaspal Subhlok

Collaboration


Dive into the Susan Hinrichs's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Corey Kosak

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Jaspal Subhlok

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Riichiro Take

Carnegie Mellon University

View shared research outputs
Researchain Logo
Decentralizing Knowledge