Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yandan Wang is active.

Publication


Featured researches published by Yandan Wang.


design automation conference | 2016

A new learning method for inference accuracy, core occupation, and performance co-optimization on TrueNorth chip

Wei Wen; Chunpeng Wu; Yandan Wang; Kent W. Nixon; Qing Wu; Mark Barnell; Hai Li; Yiran Chen

IBM TrueNorth chip uses digital spikes to perform neuromorphic computing and achieves ultrahigh execution parallelism and power efficiency. However, in TrueNorth chip, low quantization resolution of the synaptic weights and spikes significantly limits the inference (e.g., classification) accuracy of the deployed neural network model. Existing workaround, i.e., averaging the results over multiple copies instantiated in spatial and temporal domains, rapidly exhausts the hardware resources and slows down the computation. In this work, we propose a novel learning method on TrueNorth platform that constrains the random variance of each computation copy and reduces the number of needed copies. Compared to the existing learning method, our method can achieve up to 68.8% reduction of the required neuro-synaptic cores or 6.5× speedup, with even slightly improved inference accuracy.


IEEE Journal on Emerging and Selected Topics in Circuits and Systems | 2016

Leveraging Stochastic Memristor Devices in Neuromorphic Hardware Systems

Miao Hu; Yandan Wang; Wei Wen; Yu Wang; Hai Li

As the fourth basic circuit element, memristor has a unique synapse-alike feature which demonstrates great potentials in neuromorphic circuit design. However, a large gap exists between the theoretical memristor characteristics and the actual device behavior. For example, though the continuous changing in resistance state is expected in neuromorphic circuit design, it is difficult to maintain arbitrary intermediate state. In addition, the stochastic switching behaviors have been widely observed in nano-scale memristor devices. In this work, we first developed a stochastic behavior model in order to facilitate the investigation on memristor-based hardware implementation. Our modeling was based on the statistical analysis of experimental data of TiO2 device. By leveraging the stochastic behavior of memristors, a random number generator was proposed. We also presented a macro cell design composed of multiple parallel connecting memristors which can be successfully used in implementing the weight storage unit and the stochastic neuron. The designs of these fundamental components provide a feasible solution in memristor-based hardware implementation of neural networks.


ifip ieee international conference on very large scale integration | 2015

An overview on memristor crossabr based neuromorphic circuit and architecture

Zheng Li; Chenchen Liu; Yandan Wang; Bonan Yan; Chaofei Yang; Jianlei Yang; Hai Li

As technology advances, artificial intelligence becomes pervasive in society and ubiquitous in our lives, which stimulates the desire for embedded-everywhere and human-centric intelligent computation paradigm. However, conventional instruction-based computer architecture was designed for algorithmic and exact calculations. It is not suitable for handling the applications of machine learning and neural networks that usually involve a large sets of noisy and incomplete natural data. Instead, neuromorphic systems inspired by the working mechanism of human brains create promising potential. Neuromorphic systems possess a massively parallel architecture with closely coupled memory and computing. Moreover, through the sparse utilizations of hardware resources in time and space, extremely high power efficiency can be achieved. In recent years, the use of memristor technology in neuromorphic systems has attracted growing attention for its distinctive properties, such as nonvolatility, reconfigurability, and analog processing capability. In this paper, we summarize the research efforts in the development of memristor crossbar based neuromorphic design from the perspectives of device modeling, circuit, architecture, and design automation.


design automation conference | 2017

Group Scissor: Scaling Neuromorphic Computing Design to Large Neural Networks

Yandan Wang; Wei Wen; Beiye Liu; Donald M. Chiarulli; Hai Helen Li

Synapse crossbar is an elementary structure in neuromorphic computing systems (NCS). However, the limited size of crossbars and heavy routing congestion impede the NCS implementation of large neural networks. In this paper, we propose a two-step framework (namely, group scissor) to scale NCS designs to large neural networks. The first step rank clipping integrates low-rank approximation into the training to reduce total crossbar area. The second step is group connection deletion, which structurally prunes connections to reduce routing congestion between crossbars. Tested on convolutional neural networks of LeNet on MNIST database and ConvNet on CIFAR-10 database, our experiments show significant reduction of crossbar and routing area in NCS designs. Without accuracy loss, rank clipping reduces the total crossbar area to 13.62% or 51.81% in the NCS design of LeNet or ConvNet, respectively. The following group connection deletion further decreases the routing area of LeNet or ConvNet to 8.1% or 52.06%, respectively.


asia and south pacific design automation conference | 2017

Classification accuracy improvement for neuromorphic computing systems with one-level precision synapses

Yandan Wang; Wei Wen; Linghao Song; Hai Helen Li

Brain inspired neuromorphic computing has demonstrated remarkable advantages over traditional von Neumann architecture for its high energy efficiency and parallel data processing. However, the limited resolution of synaptic weights degrades system accuracy and thus impedes the use of neuromorphic systems. In this work, we propose three orthogonal methods to learn synapses with one-level precision, namely, distribution-aware quantization, quantization regularization and bias tuning, to make image classification accuracy comparable to the state-of-the-art. Experiments on both multi-layer perception and convolutional neural networks show that the accuracy drop can be well controlled within 0.19% (5.53%) for MNIST (CIFAR-10) database, compared to an ideal system without quantization.


great lakes symposium on vlsi | 2016

The Applications of NVM Technology in Hardware Security

Chaofei Yang; Beiye Liu; Yandan Wang; Yiran Chen; Hai Li; Xian Zhang; Guangyu Sun

The emerging nonvolatile memory (NVM) technologies have demonstrated great potentials in revolutionizing modern memory hierarchy because of their many promising properties: nanosecond read/write time, small cell area, non-volatility, and easy CMOS integration. It is also found that NVM devices can be leveraged to realize some hardware security solutions efficiently, such as physical unclonable function (PUF) and random number generator (RNG). In this paper, we summarize two of our works about using NVM devices to implement these hardware security features and compare them with conventional designs.


Archive | 2017

In-place Logic Obfuscation for Emerging Nonvolatile FPGAs

Yi-Chung Chen; Yandan Wang; Wei Zhang; Yiran Chen; Hai Li

To enhance system integrity of FPGA-based embedded systems on hardware design and data communication, we propose a hardware security scheme for nonvolatile resistive random access memory (RRAM) based FPGA, in which internal block RAM (BRAMs) are used for configuration and temporary data storage. The proposed scheme loads obfuscated configurations into nonvolatile BRAMs to protect design data from physical attacks and utilizes Chip DNA to enable design functionality. Moreover, in order to support run-time and remote reconfiguration even in public and insecure environment, we propose a encrypted addressing to secure communication ports with encrypted address.


international conference on computer aided design | 2016

A data locality-aware design framework for reconfigurable sparse matrix-vector multiplication kernel

Sicheng Li; Yandan Wang; Wujie Wen; Yu Wang; Yiran Chen; Hai Li

Sparse matrix-vector multiplication (SpMV) is an important computational kernel in many applications. For performance improvement, software libraries designated for SpMV computation have been introduced, e.g., MKL library for CPUs and cuSPARSE library for GPUs. However, the computational throughput of these libraries is far below the peak floating-point performance offered by hardware platforms, because the efficiency of SpMV kernel is greatly constrained by the limited memory bandwidth and irregular data access patterns. In this work, we propose a data locality-aware design framework for FPGA-based SpMV acceleration. We first include the hardware constraints in sparse matrix compression at software level to regularize the memory allocation and accesses. Moreover, a distributed architecture composed of processing elements is developed to improve the computation parallelism. We implement the reconfigurable SpMV kernel on Convey HC-2ex and conduct the evaluation by using the University of Florida sparse matrix collection. The experiments demonstrate an average computational efficiency of 48.2%, which is a lot better than those of CPU and GPU implementations. Our FPGA-based kernel has a comparable runtime as GPU, and achieves 2.1× reduction than CPU. Moreover, our design obtains substantial saving in energy consumption, say, 9.3× and 5.6× better than the implementations on CPU and GPU, respectively.


neural information processing systems | 2016

Learning Structured Sparsity in Deep Neural Networks

Wei Wen; Chunpeng Wu; Yandan Wang; Yiran Chen; Hai Li


neural information processing systems | 2017

TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning

Wei Wen; Cong Xu; Feng Yan; Chunpeng Wu; Yandan Wang; Yiran Chen; Hai Li

Collaboration


Dive into the Yandan Wang's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Wei Wen

University of Pittsburgh

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Chunpeng Wu

University of Pittsburgh

View shared research outputs
Top Co-Authors

Avatar

Cong Xu

Pennsylvania State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Beiye Liu

University of Pittsburgh

View shared research outputs
Top Co-Authors

Avatar

Chaofei Yang

University of Pittsburgh

View shared research outputs
Top Co-Authors

Avatar

Hai Helen Li

University of Pittsburgh

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge