Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Waldemar Horwat is active.

Publication


Featured researches published by Waldemar Horwat.


international symposium on computer architecture | 1987

Architecture of a message-driven processor

William J. Dally; Linda Chao; Andrew A. Chien; Soha Hassoun; Waldemar Horwat; Jon Kaplan; Paul Song; Brian Totty; D. Scott Wills

We propose a machine architecture for a high-performance processing node for a message-passing, MIMD concurrent computer. The principal mechanisms for attaining this goal are the direct execution and buffering of messages and a memory-based architecture that permits very fast context switches. Our architecture also includes a novel memory organization that permits both indexed and associative accesses and that incorporates an instruction buffer and message queue. Simulation results suggest that this architecture reduces message reception overhead by more than an order of magnitude.


programming language design and implementation | 1989

Experience with CST: programming and implementation

Waldemar Horwat; Andrew A. Chien; William J. Dally

CST is a programming language based on Smalltalk-802 that supports concurrency using locks, asynchronous messages, and distributed objects. In this paper, we describe CST: the language and its implementation. Example programs and initial programming experience with CST are described. Our implementation of CST generates native code for the J-machine, a fine-grained concurrent computer. Some compiler optimizations developed in conjunction with that implementation are also described.


international conference on computer design | 1992

The Message Driven Processor: an integrated multicomputer processing element

William J. Dally; Andrew A. Chien; J.A.S. Fiske; G. Fyler; Waldemar Horwat; John S. Keen; Richard Lethin; Michael D. Noakes; Peter R. Nuth; D.S. Wills

A description is given of the Message-Driven Processor (MDP), an integrated multicomputer node. It incorporates a 36-bit integer processor, a memory management unit, a router for a 3D mesh network, a network interface, a 4K*36-bit word static RAM (SRAM), and an ECC dynamic RAM (DRAM) controller on a single 1.1 M-transistor VLSI chip. The MDP is not specialized for a single model of computation. Instead, it incorporates efficient primitive mechanisms for communication, synchronization, and naming. These mechanisms support most proposed parallel programming models. Each processing node of the MIT J-Machine consists of an MDP with 1 Mbit of DRAM.<<ETX>>


international symposium on computer architecture | 1998

Retrospective: the J-machine

William J. Dally; Andrew A. Chien; Stuart Fiske; Waldemar Horwat; Richard Lethin; Michael D. Noakes; Peter R. Nuth; Ellen Spertus; Deborah A. Wallach; D. Scott Wills; Andrew Chang; John S. Keen

1 Computer Systems ’ Department of Computer 3 Department of Electrical Laboratory, Stanford Science, University of Illinois, and Computer Engineering, University Urbana-Champaign Georgia Institute of Technology 4 Netscape Communications 5 Equator Technologies 6 Hewlett Packard Consulting Laboratories 7 Department of Computer 8 DEC, Western Research 9 Silicon Graphics Computer Science, Mills College Laboratory Systems


Computing Systems in Engineering | 1992

The J-Machine: A fine-grain parallel computer

William J. Dally; Andrew A. Chien; R.E. Davison; J.A.S. Fiske; S. Furman; G. Fyler; D.B. Gaunce; Waldemar Horwat; S. Kaneshiro; John S. Keen; Richard Lethin; Michael D. Noakes; Peter R. Nuth; Ellen Spertus; Brian Totty; Deborah A. Wallach; D.S. Wills

Abstract Most modern computers, whether parallel or sequential, are coarse grained. They are composed of physically large nodes with tens of megabytes of memory. Only a small fraction of the silicon area in the machine is devoted to computation. By increasing the ratio of computation area to memory area, fine-grain computers offer the potential of improving cost/performance by several orders of magnitude. To efficiently operate at such a fine grain, however, a machine must provide mechanisms that permit rapid access to global data and fast interaction between nodes. The MIT J-Machine is a fine-grain concurrent computer that provides low-overhead mechanisms for parallel computing. Prototype J-Machines have been operational since July 1991. The J-Machine communication mechanism permits a node to send a message to any other node in the machine in μ s. On message arrival, a task is created and dispatched in μ s. A translation mechanism supports a global virtual address space. These mechanisms efficiently support most proposed models of concurrent computation and allow parallelism to be exploited at a grain size of 10 operations. The hardware is an ensemble of up to 65,536 nodes each containing a 36-bit processor, 4K 36-bit words of on-chip memory, 256K words of DRAM and a router. The nodes are connected by a high-speed three-dimensional mesh network.


ifip congress | 1989

The J-Machine: A Fine Grain Concurrent Computer

William J. Dally; Andrew A. Chien; Stuart Fiske; Waldemar Horwat; John S. Keen


Archive | 1988

Message-driven processor in a concurrent computer

William J. Dally; Andrew A. Chien; Waldemar Horwat; Stuart Fiske


Archive | 1988

Message-Driven Processor Architecture, Version 11

William J. Dally; Andrew A. Chien; Stuart Fiske; Waldemar Horwat; John S. Keen; Peter R. Nuth; Jerry Larivee; Brian Totty


Research directions in concurrent object-oriented programming | 1993

COSMOS: an operating system for a fine-grain concurrent computer

Waldemar Horwat; Brian Totty; William J. Dally


Archive | 1989

J-machine: A fine-grain concurrent computer

William J. Dally; Andrew A. Chien; Stuart Fiske; Waldemar Horwat; John S. Keen

Collaboration


Dive into the Waldemar Horwat's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

John S. Keen

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Stuart Fiske

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Brian Totty

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Peter R. Nuth

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Michael D. Noakes

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Richard Lethin

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

D. Scott Wills

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

D.S. Wills

Massachusetts Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge