Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Vivienne Sze is active.

Publication


Featured researches published by Vivienne Sze.


international symposium on computer architecture | 2016

Eyeriss: a spatial architecture for energy-efficient dataflow for convolutional neural networks

Yu-Hsin Chen; Joel S. Emer; Vivienne Sze

Deep convolutional neural networks (CNNs) are widely used in modern AI systems for their superior accuracy but at the cost of high computational complexity. The complexity comes from the need to simultaneously process hundreds of filters and channels in the high-dimensional convolutions, which involve a significant amount of data movement. Although highly-parallel compute paradigms, such as SIMD/SIMT, effectively address the computation requirement to achieve high throughput, energy consumption still remains high as data movement can be more expensive than computation. Accordingly, finding a dataflow that supports parallel processing with minimal data movement cost is crucial to achieving energy-efficient CNN processing without compromising accuracy.n In this paper, we present a novel dataflow, called row-stationary (RS), that minimizes data movement energy consumption on a spatial architecture. This is realized by exploiting local data reuse of filter weights and feature map pixels, i.e., activations, in the high-dimensional convolutions, and minimizing data movement of partial sum accumulations. Unlike dataflows used in existing designs, which only reduce certain types of data movement, the proposed RS dataflow can adapt to different CNN shape configurations and reduces all types of data movement through maximally utilizing the processing engine (PE) local storage, direct inter-PE communication and spatial parallelism. To evaluate the energy efficiency of the different dataflows, we propose an analysis framework that compares energy cost under the same hardware area and processing parallelism constraints. Experiments using the CNN configurations of AlexNet show that the proposed RS dataflow is more energy efficient than existing dataflows in both convolutional (1.4× to 2.5×) and fully-connected layers (at least 1.3× for batch size larger than 16). The RS dataflow has also been demonstrated on a fabricated chip, which verifies our energy analysis.


IEEE Journal of Solid-state Circuits | 2017

Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural Networks

Yu-Hsin Chen; Tushar Krishna; Joel S. Emer; Vivienne Sze

Deep learning using convolutional neural networks (CNN) gives state-of-the-art accuracy on many computer vision tasks (e.g. object detection, recognition, segmentation). Convolutions account for over 90% of the processing in CNNs for both inference/testing and training, and fully convolutional networks are increasingly being used. To achieve state-of-the-art accuracy requires CNNs with not only a larger number of layers, but also millions of filters weights, and varying shapes (i.e. filter sizes, number of filters, number of channels) as shown in Fig. 14.5.1. For instance, AlexNet [1] uses 2.3 million weights (4.6MB of storage) and requires 666 million MACs per 227×227 image (13kMACs/pixel). VGG16 [2] uses 14.7 million weights (29.4MB of storage) and requires 15.3 billion MACs per 224×224 image (306kMACs/pixel). The large number of filter weights and channels results in substantial data movement, which consumes significant energy.


arXiv: Computer Vision and Pattern Recognition | 2017

Efficient Processing of Deep Neural Networks: A Tutorial and Survey

Vivienne Sze; Yu-Hsin Chen; Tien-Ju Yang; Joel S. Emer

Deep neural networks (DNNs) are currently widely used for many artificial intelligence (AI) applications including computer vision, speech recognition, and robotics. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Accordingly, techniques that enable efficient processing of DNNs to improve energy efficiency and throughput without sacrificing application accuracy or increasing hardware cost are critical to the wide deployment of DNNs in AI systems. This article aims to provide a comprehensive tutorial and survey about the recent advances toward the goal of enabling efficient processing of DNNs. Specifically, it will provide an overview of DNNs, discuss various hardware platforms and architectures that support DNNs, and highlight key trends in reducing the computation cost of DNNs either solely via hardware design changes or via joint hardware design and DNN algorithm changes. It will also summarize various development resources that enable researchers and practitioners to quickly get started in this field, and highlight important benchmarking metrics and design considerations that should be used for evaluating the rapidly growing number of DNN hardware designs, optionally including algorithmic codesigns, being proposed in academia and industry. The reader will take away the following concepts from this article: understand the key design considerations for DNNs; be able to evaluate different DNN hardware implementations with benchmarks and comparison metrics; understand the tradeoffs between various hardware architectures and platforms; be able to evaluate the utility of various DNN design techniques for efficient processing; and understand recent implementation trends and opportunities.


computer vision and pattern recognition | 2017

Designing Energy-Efficient Convolutional Neural Networks Using Energy-Aware Pruning

Tien-Ju Yang; Yu-Hsin Chen; Vivienne Sze

Deep convolutional neural networks (CNNs) are indispensable to state-of-the-art computer vision algorithms. However, they are still rarely deployed on battery-powered mobile devices, such as smartphones and wearable gadgets, where vision algorithms can enable many revolutionary real-world applications. The key limiting factor is the high energy consumption of CNN processing due to its high computational complexity. While there are many previous efforts that try to reduce the CNN model size or the amount of computation, we find that they do not necessarily result in lower energy consumption. Therefore, these targets do not serve as a good metric for energy cost estimation. To close the gap between CNN design and energy consumption optimization, we propose an energy-aware pruning algorithm for CNNs that directly uses the energy consumption of a CNN to guide the pruning process. The energy estimation methodology uses parameters extrapolated from actual hardware measurements. The proposed layer-by-layer pruning algorithm also prunes more aggressively than previously proposed pruning methods by minimizing the error in the output feature maps instead of the filter weights. For each layer, the weights are first pruned and then locally fine-tuned with a closed-form least-square solution to quickly restore the accuracy. After all layers are pruned, the entire network is globally fine-tuned using back-propagation. With the proposed pruning method, the energy consumption of AlexNet and GoogLeNet is reduced by 3.7x and 1.6x, respectively, with less than 1% top-5 accuracy loss. We also show that reducing the number of target classes in AlexNet greatly decreases the number of weights, but has a limited impact on energy consumption.


IEEE Micro | 2017

Using Dataflow to Optimize Energy Efficiency of Deep Neural Network Accelerators

Yu-Hsin Chen; Joel S. Emer; Vivienne Sze

The authors demonstrate the key role dataflows play in the optimization of energy efficiency for deep neural network (DNN) accelerators. By introducing a systematic approach to analyze the problem and a new dataflow, called Row-Stationary, which is up to 2.5 times more energy efficient than existing dataflows in processing a state-of-the-art DNN, this work provides guidelines for future DNN accelerator designs.


arXiv: Computer Vision and Pattern Recognition | 2018

NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications.

Tien-Ju Yang; Andrew Howard; Bo Chen; Xiao Zhang; Alec Go; Vivienne Sze; Hartwig Adam


asilomar conference on signals, systems and computers | 2017

A method to estimate the energy consumption of deep neural networks

Tien-Ju Yang; Yu-Hsin Chen; Joel S. Emer; Vivienne Sze


IEEE Micro | 2017

Eyeriss: A Spatial Architecture for Energy-Efficient Dataflow for Convolutional Neural Networks

Yu-Hsin Chen; Joel S. Emer; Vivienne Sze


arXiv: Distributed, Parallel, and Cluster Computing | 2018

Eyeriss v2: A Flexible and High-Performance Accelerator for Emerging Deep Neural Networks.

Yu-Hsin Chen; Joel S. Emer; Vivienne Sze


custom integrated circuits conference | 2017

Hardware for machine learning: Challenges and opportunities

Vivienne Sze; Yu-Hsin Chen; Joel Einer; Amr Suleiman; Zhengdong Zhang

Collaboration


Dive into the Vivienne Sze's collaboration.

Top Co-Authors

Avatar

Yu-Hsin Chen

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Joel S. Emer

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Tien-Ju Yang

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Amr Suleiman

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Zhengdong Zhang

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Joel Einer

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Tushar Krishna

Georgia Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge