Parami Wijesinghe
Purdue University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Parami Wijesinghe.
Scientific Reports | 2016
Abhronil Sengupta; Priyadarshini Panda; Parami Wijesinghe; Yusung Kim; Kaushik Roy
Brain-inspired computing architectures attempt to mimic the computations performed in the neurons and the synapses in the human brain in order to achieve its efficiency in learning and cognitive tasks. In this work, we demonstrate the mapping of the probabilistic spiking nature of pyramidal neurons in the cortex to the stochastic switching behavior of a Magnetic Tunnel Junction in presence of thermal noise. We present results to illustrate the efficiency of neuromorphic systems based on such probabilistic neurons for pattern recognition tasks in presence of lateral inhibition and homeostasis. Such stochastic MTJ neurons can also potentially provide a direct mapping to the probabilistic computing elements in Belief Networks for performing regenerative tasks.
IEEE Transactions on Nanotechnology | 2015
Kon-Woo Kwon; Xuanyao Fong; Parami Wijesinghe; Priyadarshini Panda; Kaushik Roy
In spin-transfer torque magnetic random access memory (STT-MRAM), retention-, write-, and read-failures negatively impact the memory yield and density. In this paper, we jointly consider device-circuit-architecture layers to implement high-density STT-MRAM array while meeting the target yield requirement. Different types of magnetic tunnel junctions are considered at the device level, and error correcting codes (ECCs) in conjunction with invert-coding are employed as an architectural solution. Through cross-layer interactions, we present a design methodology to optimize bit-cell area while satisfying the target yield and energy consumption under process variation. Furthermore, we explore the use of invert-coding along with ECC in order to achieve higher memory density than that obtained using ECC alone. Our proposed technique can improve memory density further by proper selection of thermal stability factor based upon two observations: 1) invert-coding can fix multiple write/read failures with small storage overhead and 2) as thermal stability factor increases, retention-failure probability exponentially decreases, and thus, simple ECC is good enough for retention failure correction.
design, automation, and test in europe | 2016
Gopalakrishnan Srinivasan; Parami Wijesinghe; Syed Shakib Sarwar; Akhilesh Jaiswal; Kaushik Roy
Multilayered artificial neural networks have found widespread utility in classification and recognition applications. The scale and complexity of such networks together with the inadequacies of general purpose computing platforms have led to a significant interest in the development of efficient hardware implementations. In this work, we focus on designing energy-efficient on-chip storage for the synaptic weights, motivated primarily by the observation that the number of synapses is orders of magnitude larger than the number of neurons. Typical digital CMOS implementations of such large-scale networks are power hungry. In order to minimize the power consumption, the digital neurons could be operated reliably at scaled voltages by reducing the clock frequency. On the contrary, the on-chip synaptic storage designed using a conventional 6T SRAM is susceptible to bitcell failures at reduced voltages. However, the intrinsic error resiliency of neural networks to small synaptic weight perturbations enables us to scale the operating voltage of the 6T SRAM. Our analysis on a widely used digit recognition dataset indicates that the voltage can be scaled by 200 mV from the nominal operating voltage (950 mV) for practically no loss (less than 0.5%) in accuracy (22 nm predictive technology). Scaling beyond that causes substantial performance degradation owing to increased probability of failures in the MSBs of the synaptic weights. We, therefore propose a significance driven hybrid 8T-6T SRAM, wherein the sensitive MSBs are stored in 8T bitcells that are robust at scaled voltages due to decoupled read and write paths. In an effort to further minimize the area penalty, we present a synaptic-sensitivity driven hybrid memory architecture consisting of multiple 8T-6T SRAM banks. Our circuit to system-level simulation framework shows that the proposed synaptic-sensitivity driven architecture provides a 30.91% reduction in the memory access power with a 10.41% area overhead, for less than 1% loss in the classification accuracy.
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems | 2017
Priyadarshini Panda; Aayush Ankit; Parami Wijesinghe; Kaushik Roy
Machine-learning algorithms have shown outstanding image recognition/classification performance for computer vision applications. However, the compute and energy requirement for implementing such classifier models for large-scale problems is quite high. In this paper, we propose feature driven selective classification (FALCON) inspired by the biological visual attention mechanism in the brain to optimize the energy-efficiency of machine-learning classifiers. We use the consensus in the characteristic features (color/texture) across images in a dataset to decompose the original classification problem and construct a tree of classifiers (nodes) with a generic-to-specific transition in the classification hierarchy. The initial nodes of the tree separate the instances based on feature information and selectively enable the latter nodes to perform object specific classification. The proposed methodology allows selective activation of only those branches and nodes of the classification tree that are relevant to the input while keeping the remaining nodes idle. Additionally, we propose a programmable and scalable neuromorphic engine (NeuE) that utilizes arrays of specialized neural computational elements to execute the FALCON-based classifier models for diverse datasets. The structure of FALCON facilitates the reuse of nodes while scaling up from small classification problems to larger ones thus allowing us to construct classifier implementations that are significantly more efficient. We evaluate our approach for a 12-object classification task on the Caltech101 dataset and ten-object task on CIFAR-10 dataset by constructing FALCON models on the NeuE platform in 45-nm technology. Our results demonstrate up to
Scientific Reports | 2018
Parami Wijesinghe; Chamika M. Liyanagedera; Kaushik Roy
3.66\boldsymbol \times
international symposium on neural networks | 2017
Chamika M. Liyanagedera; Parami Wijesinghe; Akhilesh Jaiswal; Kaushik Roy
improvement in energy-efficiency for no loss in output quality, and even higher improvements of up to
design, automation, and test in europe | 2017
Parami Wijesinghe; Chamika M. Liyanagedera; Kaushik Roy
5.91\boldsymbol \times
Scientific Reports | 2017
Abhronil Sengupta; Priyadarshini Panda; Parami Wijesinghe; Yusung Kim; Kaushik Roy
with 3.9% accuracy loss compared to an optimized baseline network. In addition, FALCON shows an improvement in training time of up to
arXiv: Emerging Technologies | 2018
Parami Wijesinghe; Aayush Ankit; Abhronil Sengupta; Kaushik Roy
1.96\boldsymbol \times
IEEE Journal on Emerging and Selected Topics in Circuits and Systems | 2018
Syed Shakib Sarwar; Gopalakrishnan Srinivasan; Bing Han; Parami Wijesinghe; Akhilesh Jaiswal; Priyadarshini Panda; Anand Raghunathan; Kaushik Roy
as compared to the traditional classification approach.