Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Alain J. Martin is active.

Publication


Featured researches published by Alain J. Martin.


Distributed Computing | 1986

Compiling Communicating Processes into Delay-Insensitive VLSI Circuits

Alain J. Martin

A method is described for compiling computations described as a set of communicating processes into VLSI circuits. The circuits obtained are delay-insensitive, i.e., their correct operation is independent of any assumption on delays in operators and wires, except that the delays are finite. They are also correct by construction since they are derived by a series of semantics-preserving transformation.


conference on advanced research in vlsi | 1990

The limitations to delay-insensitivity in asynchronous circuits

Alain J. Martin

Asynchronous techniques —that is, techniques that do not use clocks to implement sequencing— are currently attracting considerable interest for digital VLSI circuit design, in particular when the circuits produced are delay-insensitive (DI). A digital circuit is DI when its correct operation is independent of the delays in operators and in the wires connecting the operators, except that the delays are finite and positive.


conference on advanced research in vlsi | 1997

The design of an asynchronous MIPS R3000 microprocessor

Alain J. Martin; Andrew Lines; Rajit Manohar; Mika Nyström; Paul I. Pénzes; Robert Southworth; Uri Cummings; Tak Kwan Lee

The design of an asynchronous clone of a MIPS R3000 microprocessor is presented. In 0.6 /spl mu/m CMOS, we expect performance close to 280 MIPS, for a power consumption of 7 W. The paper describes the structure of a high-performance asynchronous pipeline, in particular precise exceptions, pipelined caches, arithmetic, and registers, and the circuit techniques developed to achieve high throughput.


Proceedings of the IEEE | 2006

Asynchronous Techniques for System-on-Chip Design

Alain J. Martin; Mika Nyström

SoC design will require asynchronous techniques as the large parameter variations across the chip will make it impossible to control delays in clock networks and other global signals efficiently. Initially, SoCs will be globally asynchronous and locally synchronous (GALS). But the complexity of the numerous asynchronous/synchronous interfaces required in a GALS will eventually lead to entirely asynchronous solutions. This paper introduces the main design principles, methods, and building blocks for asynchronous VLSI systems, with an emphasis on communication and synchronization. Asynchronous circuits with the only delay assumption of isochronic forks are called quasi-delay-insensitive (QDI). QDI is used in the paper as the basis for asynchronous logic. The paper discusses asynchronous handshake protocols for communication and the notion of validity/neutrality tests, and completion tree. Basic building blocks for sequencing, storage, function evaluation, and buses are described, and two alternative methods for the implementation of an arbitrary computation are explained. Issues of arbitration, and synchronization play an important role in complex distributed systems and especially in GALS. The two main asynchronous/synchronous interfaces needed in GALS-one based on synchronizer, the other on stoppable clock-are described and analyzed.


hypercube concurrent computers and applications | 1988

The architecture and programming of the Ametek series 2010 multicomputer

Charles L. Seitz; William C. Athas; Charles M. Flaig; Alain J. Martin; Jakov Seizovic; Craig S. Steele; Wen-King Su

During the period following the completion of the Cosmic Cube experiment [1], and while commercial descendants of this first-generation multicomputer (message-passing concurrent computer) were spreading through a community that includes many of the attendees of this conference, members of our research group were developing a set of ideas about the physical design and programming for the second generation of medium-grain multicomputers. Our principal goal was to improve by as much as two orders of magnitude the relationship between message-passing and computing performance, and also to make the topology of the message-passing network practically invisible. Decreasing the communication latency relative to instruction execution times extends the application span of multicomputers from easily partitioned and distributed problems (eg, matrix computations, PDE solvers, finite element analysis, finite difference methods, distant or local field many-body problems, FFTs, ray tracing, distributed simulation of systems composed of loosely coupled physical processes) to computing problems characterized by “high flux” [2] or relatively fine-grain concurrent formulations [3, 4] (eg, searching, sorting, concurrent data structures, graph problems, signal processing, image processing, and distributed simulation of systems composed of many tightly coupled physical processes). Such applications place heavy demands on the message-passing network for high bandwidth, low latency, and non-local communication. Decreased message latency also improves the efficiency of the class of applications that have been developed on first-generation systems, and the insensitivity of message latency to process placement simplifies the concurrent formulation of application programs. Our other goals included a streamlined and easily layered set of message primitives, a node operating system based on a reactive programming model, open interfaces for accelerators and peripheral devices, and node performance improvements that could be achieved economically by using the same technology employed in contemporary workstation computers. By the autumn of 1986, these ideas had become sufficiently developed, molded together, and tested through simulation to be regarded as a complete architectural design. We were fortunate that the Ametek Computer Research Division was ready and willing to work with us to develop this system as a commercial product. The Ametek Series 2010 multicomputer is the result of this joint effort.


Formal Methods in System Design | 1992

Asynchronous datapaths and the design of an asynchronous adder

Alain J. Martin

This article presents a general method for designing delay-insensitive datapath circuits. Its emphasis is on the formal derivation of a circuit from its specification. We discuss the properties required in a code that is used to transmit data asynchronously, and we introduce such a code. We introduce a general method (in the form of a theorem) for distributing the evaluation of a function over a number of concurrent cells. This method requires that the code be distributive. We apply the method to the familiar example of a ripple-carry adder, and we give a CMOS implementation of the adder.


symposium on asynchronous circuits and systems | 2003

The Lutonium: a sub-nanojoule asynchronous 8051 microcontroller

Alain J. Martin; Mika Nyström; Karl Papadantonakis; Paul I. Pénzes; Piyush Prakash; Catherine G. Wong; Jonathan Chang; Kevin S. Ko; Benjamin N. Lee; Elaine Ou; Jim Pugh; Eino-Ville Talvala; James T. Tong; Ahmet Tura

We describe the Lutonium, an asynchronous 8051 microcontroller designed for low Et/sup 2/. In 0.18 /spl mu/m CMOS, at nominal 1.8 V, we expect a performance of 0.5 nJ per instruction at 200 MIPS. At 0.5 V, we expect 4 MIPS and 40 pJ/instruction, corresponding to 25,000 MIPS/Watt. We describe the structure of a fine-grain pipeline optimized for Et/sup 2/ efficiency, some of the peripherals implementation, and the advantages of an asynchronous implementation of a deep-sleep mechanism.


mathematics of program construction | 1998

Slack Elasticity in Concurrent Computing

Rajit Manohar; Alain J. Martin

We present conditions under which we can modify the slack of a channel in a distributed computation without changing its behavior. These results can be used to modify the degree of pipelining in an asynchronous system. The generality of the result shows the wide variety of pipelining alternatives presented to the designer of a concurrent system. We give examples of program transformations which can be used in the design of concurrent systems whose correctness depends on the conditions presented.


Power aware computing | 2002

ET 2 : a metric for time and energy efficiency of computation

Alain J. Martin; Mika Nyström; Paul I. Pénzes

We investigate an efficiency metric for VLSI computation that includes energy. E, and time, t, in the form Et2. We apply the metric to CMOS circuits operating outside velocity saturation when energy and delay can be exchanged by adjusting the supply voltage; we prove that under these assumptions, optimal Et2 implies optimal energy and delay. We give experimental and simulation evidences of the range and limits of the assumptions. We derive several results about sequential, parallel, and pipelined computations optimized for Et2, including a result about the optimal length of a pipeline.We discuss transistor sizing for optimal Et2 and show that, for fixed, nonzero execution rates, the optimum is achieved when the sum of the transistor-gate capacitances is twice the sum of the parasitic capacitances--not for minimum transistor sizes. We derive an approximation for Etn (for arbitrary n) of an optimally sized system that can be computed without actually sizing the transistors; we show that this approximation is accurate. We prove that when multiple, adjustable supply voltages are allowed, the optimal Et2 for the sequential composition of components is achieved when the supply voltages are adjusted so that the components consume equal power. Finally, we give rules for computing the Et2 of the sequential and parallel compositions of systems, when the Et2 of the components are known.


Information Processing Letters | 2001

Towards an energy complexity of computation

Alain J. Martin

Abstract Energy consumption is becoming a critical complexity parameter along with time (delay) in the design and optimization of algorithms at both the hardware and software levels. This paper proposes that a new complexity measure including energy E and time t in the form of the expression E×t 2 be used as the measure of the efficiency of a computation. We prove that the metric is optimal. As an example, a new result concerning the optimal length of a pipeline is derived.

Collaboration


Dive into the Alain J. Martin's collaboration.

Top Co-Authors

Avatar

Mika Nyström

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Catherine G. Wong

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Paul I. Pénzes

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Andrew Lines

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Mika Nystroem

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Sean Keller

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Uri Cummings

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Steven M. Burns

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jose A. Tierno

California Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge