Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Rathinakumar Appuswamy is active.

Publication


Featured researches published by Rathinakumar Appuswamy.


Science | 2014

A million spiking-neuron integrated circuit with a scalable communication network and interface

Paul A. Merolla; John V. Arthur; Rodrigo Alvarez-Icaza; Andrew S. Cassidy; Jun Sawada; Filipp Akopyan; Bryan L. Jackson; Nabil Imam; Chen Guo; Yutaka Nakamura; Bernard Brezzo; Ivan Vo; Steven K. Esser; Rathinakumar Appuswamy; Brian Taba; Arnon Amir; Myron Flickner; William P. Risk; Rajit Manohar; Dharmendra S. Modha

Modeling computer chips on real brains Computers are nowhere near as versatile as our own brains. Merolla et al. applied our present knowledge of the structure and function of the brain to design a new computer chip that uses the same wiring rules and architecture. The flexible, scalable chip operated efficiently in real time, while using very little power. Science, this issue p. 668 A large-scale computer chip mimics many features of a real brain. Inspired by the brain’s structure, we have developed an efficient, scalable, and flexible non–von Neumann architecture that leverages contemporary silicon technology. To demonstrate, we built a 5.4-billion-transistor chip with 4096 neurosynaptic cores interconnected via an intrachip network that integrates 1 million programmable spiking neurons and 256 million configurable synapses. Chips can be tiled in two dimensions via an interchip communication interface, seamlessly scaling the architecture to a cortexlike sheet of arbitrary size. The architecture is well suited to many applications that use complex neural networks in real time, for example, multiobject detection and classification. With 400-pixel-by-240-pixel video input at 30 frames per second, the chip consumes 63 milliwatts.


Proceedings of the National Academy of Sciences of the United States of America | 2016

Convolutional networks for fast, energy-efficient neuromorphic computing

Steven K. Esser; Paul A. Merolla; John V. Arthur; Andrew S. Cassidy; Rathinakumar Appuswamy; Alexander Andreopoulos; David J. Berg; Jeffrey L. McKinstry; Timothy Melano; R Davis; Carmelo di Nolfo; Pallab Datta; Arnon Amir; Brian Taba; Myron Flickner; Dharmendra S. Modha

Significance Brain-inspired computing seeks to develop new technologies that solve real-world problems while remaining grounded in the physical requirements of energy, speed, and size. Meeting these challenges requires high-performing algorithms that are capable of running on efficient hardware. Here, we adapt deep convolutional neural networks, which are today’s state-of-the-art approach for machine perception in many domains, to perform classification tasks on neuromorphic hardware, which is today’s most efficient platform for running neural networks. Using our approach, we demonstrate near state-of-the-art accuracy on eight datasets, while running at between 1,200 and 2,600 frames/s and using between 25 and 275 mW. Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that (i) approach state-of-the-art classification accuracy across eight standard datasets encompassing vision and speech, (ii) perform inference while preserving the hardware’s underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1,200 and 2,600 frames/s and using between 25 and 275 mW (effectively >6,000 frames/s per Watt), and (iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. This approach allows the algorithmic power of deep learning to be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer.


international symposium on neural networks | 2013

Cognitive computing systems: Algorithms and applications for networks of neurosynaptic cores

Steven K. Esser; Alexander Andreopoulos; Rathinakumar Appuswamy; Pallab Datta; Davis; Arnon Amir; John V. Arthur; Andrew S. Cassidy; Myron Flickner; Paul Merolla; Shyamal Chandra; Nicola Basilico; Stefano Carpin; Tom Zimmerman; Frank Zee; Rodrigo Alvarez-Icaza; Jeffrey A. Kusnitz; Theodore M. Wong; William P. Risk; Emmett McQuinn; Tapan Kumar Nayak; Raghavendra Singh; Dharmendra S. Modha

Marching along the DARPA SyNAPSE roadmap, IBM unveils a trilogy of innovations towards the TrueNorth cognitive computing system inspired by the brains function and efficiency. The non-von Neumann nature of the TrueNorth architecture necessitates a novel approach to efficient system design. To this end, we have developed a set of abstractions, algorithms, and applications that are natively efficient for TrueNorth. First, we developed repeatedly-used abstractions that span neural codes (such as binary, rate, population, and time-to-spike), long-range connectivity, and short-range connectivity. Second, we implemented ten algorithms that include convolution networks, spectral content estimators, liquid state machines, restricted Boltzmann machines, hidden Markov models, looming detection, temporal pattern matching, and various classifiers. Third, we demonstrate seven applications that include speaker recognition, music composer recognition, digit recognition, sequence prediction, collision avoidance, optical flow, and eye detection. Our results showcase the parallelism, versatility, rich connectivity, spatio-temporality, and multi-modality of the TrueNorth architecture as well as compositionality of the corelet programming paradigm and the flexibility of the underlying neuron model.


ieee international conference on high performance computing data and analytics | 2014

Real-time scalable cortical computing at 46 giga-synaptic OPS/watt with ~100× speedup in time-to-solution and ~100,000× reduction in energy-to-solution

Andrew S. Cassidy; Rodrigo Alvarez-Icaza; Filipp Akopyan; Jun Sawada; John V. Arthur; Paul A. Merolla; Pallab Datta; Marc Gonzalez Tallada; Brian Taba; Alexander Andreopoulos; Arnon Amir; Steven K. Esser; Jeff Kusnitz; Rathinakumar Appuswamy; Chuck Haymes; Bernard Brezzo; Roger Moussalli; Ralph Bellofatto; Christian W. Baks; Michael Mastro; Kai Schleupen; Charles Edwin Cox; Ken Inoue; Steven Edward Millman; Nabil Imam; Emmett McQuinn; Yutaka Nakamura; Ivan Vo; Chen Guok; Don Nguyen

Drawing on neuroscience, we have developed a parallel, event-driven kernel for neurosynaptic computation, that is efficient with respect to computation, memory, and communication. Building on the previously demonstrated highly optimized software expression of the kernel, here, we demonstrate True North, a co-designed silicon expression of the kernel. True North achieves five orders of magnitude reduction in energy to-solution and two orders of magnitude speedup in time-to solution, when running computer vision applications and complex recurrent neural network simulations. Breaking path with the von Neumann architecture, True North is a 4,096 core, 1 million neuron, and 256 million synapse brain-inspired neurosynaptic processor, that consumes 65mW of power running at real-time and delivers performance of 46 Giga-Synaptic OPS/Watt. We demonstrate seamless tiling of True North chips into arrays, forming a foundation for cortex-like scalability. True Norths unprecedented time-to-solution, energy-to-solution, size, scalability, and performance combined with the underlying flexibility of the kernel enable a broad range of cognitive applications.


allerton conference on communication, control, and computing | 2008

Network coding for computing

Rathinakumar Appuswamy; Massimo Franceschetti; Nikhil Karamchandani; Kenneth Zeger

The following network computation problem is considered. A set of source nodes in an acyclic network generates independent messages and a single receiver node computes a function f of the messages. The objective is to characterize the maximum number of times f can be computed per network usage. The network coding problem for a single receiver network is a special case of the network computation problem (taking f to be the identity map) in which all of the source messages must be reproduced at the receiver. For network coding with a single receiver, routing is known to be rate-optimal and achieves the network min-cut upper bound. We give a generalized min-cut upper bound for the network computation problem. We show that the bound reduces to the usual network min-cut when f is the identity and the bound is tight for the computation of ldquodivisible functionsrdquo over ldquotree networksrdquo. We also show that the bound is not tight in general.


IEEE Transactions on Information Theory | 2013

Linear Codes, Target Function Classes, and Network Computing Capacity

Rathinakumar Appuswamy; Massimo Franceschetti; Nikhil Karamchandani; Kenneth Zeger

We study the use of linear codes for network computing in single-receiver networks with various classes of target functions of the source messages. Such classes include reducible, semi-injective, and linear target functions over finite fields. Computing capacity bounds and achievability are given with respect to these target function classes for network codes that use routing, linear coding, or nonlinear coding.


IEEE Transactions on Information Theory | 2014

Computing Linear Functions by Linear Coding Over Networks

Rathinakumar Appuswamy; Massimo Franceschetti

We consider the scenario in which a set of sources generates messages in a network and a receiver node demands an arbitrary linear function of these messages. We formulate an algebraic test to determine whether an arbitrary network can compute linear functions using linear codes. We identify a class of linear functions that can be computed using linear codes in every network that satisfies a natural cut-based condition. Conversely, for another class of linear functions, we show that the cut-based condition does not guarantee the existence of a linear coding solution. For linear functions over the binary field, the two classes are complements of each other.


international symposium on information theory | 2009

Network computing capacity for the reverse butterfly network

Rathinakumar Appuswamy; Massimo Franceschetti; Nikhil Karamchandani; Kenneth Zeger

We study the computation of the arithmetic sum of the q-ary source messages in the reverse butterfly network. Specifically, we characterize the maximum rate at which the message sum can be computed at the receiver and demonstrate that linear coding is suboptimal.


IEEE Transactions on Information Theory | 2011

Time and Energy Complexity of Function Computation Over Networks

Nikhil Karamchandani; Rathinakumar Appuswamy; Massimo Franceschetti

This paper considers the following network computation problem: n nodes are placed on a √n × √n grid, each node is connected to every other node within distance r(n) of itself, and it is assigned an arbitrary input bit. Nodes communicate with their neighbors and a designated sink node computes a function f of the input bits, where f is either the identity or a symmetric function. We first consider a model where links are interference and noise-free, suitable for modeling wired networks. Then, we consider a model suitable for wireless networks. Due to interference, only nodes which do not share neighbors are allowed to transmit simultaneously, and when a node transmits a bit, all of its neighbors receive an independent noisy copy of the bit. We present lower bounds on the minimum number of transmissions and on the minimum number of time slots required to compute f. We also describe efficient schemes that match both of these lower bounds up to a constant factor and are thus jointly (near) optimal with respect to the number of transmissions and the number of time slots required for computation. At the end of the paper, we extend results on symmetric functions to general network topologies, and obtain a corollary that answers an open question posed by El Gamal in 1987 regarding the computation of the parity function over ring and tree networks.


IEEE Wireless Communications | 2006

QoS guarantee and provisioning for realtime digital video over mobile ad hoc cdma networks with cross-layer design

Qi Qu; Rathinakumar Appuswamy; Yee Sin Chan

In this article we investigate the trade-offs and the constraints for multimedia over mobile ad hoc CDMA networks, and propose a cross-layer distributed power control and scheduling protocol to resolve those trade-offs and constraints in order to provide high-quality video over wireless ad hoc CDMA networks. In particular, a distributed power control and scheduling protocol is proposed to control the incurred delay of video streaming over multihop wireless ad hoc networks, as well as the multiple access interference (MAI). We also investigate the impacts of Doppler spread and noisy channel estimates upon the end-to-end video quality, and provide a relatively robust system which employs a combination of power control and coding/interleaving to combat the effects of Doppler spread by exploiting the increased time diversity when the Doppler spread becomes large. Thus, more robust end-to-end video quality can be achieved over a wide range of channel conditions

Collaboration


Dive into the Rathinakumar Appuswamy's collaboration.

Researchain Logo
Decentralizing Knowledge