Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Patrick Judd is active.

Publication


Featured researches published by Patrick Judd.


international symposium on computer architecture | 2016

Cnvlutin: ineffectual-neuron-free deep neural network computing

Jorge Albericio; Patrick Judd; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger; Andreas Moshovos

This work observes that a large fraction of the computations performed by Deep Neural Networks (DNNs) are intrinsically ineffectual as they involve a multiplication where one of the inputs is zero. This observation motivates Cnvolutin (CNV), a value-based approach to hardware acceleration that eliminates most of these ineffectual operations, improving performance and energy over a state-of-the-art accelerator with no accuracy loss. CNV uses hierarchical data-parallel units, allowing groups of lanes to proceed mostly independently enabling them to skip over the ineffectual computations. A co-designed data storage format encodes the computation elimination decisions taking them off the critical path while avoiding control divergence in the data parallel units. Combined, the units and the data storage format result in a data-parallel architecture that maintains wide, aligned accesses to its memory hierarchy and that keeps its data lanes busy. By loosening the ineffectual computation identification criterion, CNV enables further performance and energy efficiency improvements, and more so if a loss in accuracy is acceptable. Experimental measurements over a set of state-of-the-art DNNs for image classification show that CNV improves performance over a state-of-the-art accelerator from 1.24× to 1.55× and by 1.37× on average without any loss in accuracy by removing zero-valued operand multiplications alone. While CNV incurs an area overhead of 4.49%, it improves overall EDP (Energy Delay Product) and ED2P (Energy Delay Squared Product) on average by 1.47× and 2.01×, respectively. The average performance improvements increase to 1.52× without any loss in accuracy with a broader ineffectual identification policy. Further improvements are demonstrated with a loss in accuracy.


international symposium on microarchitecture | 2017

Bit-pragmatic deep neural network computing

Jorge Albericio; Alberto Delmas; Patrick Judd; Sayeh Sharify; Gerard O'Leary; Roman Genov; Andreas Moshovos

Deep Neural Networks expose a high degree of parallelism, making them amenable to highly data parallel architectures. However, data-parallel architectures often accept inefficiency in individual computations for the sake of overall efficiency. We show that on average, activation values of convolutional layers during inference in modern Deep Convolutional Neural Networks (CNNs) contain 92% zero bits. Processing these zero bits entails ineffectual computations that could be skipped. We propose Pragmatic (PRA), a massively data-parallel architecture that eliminates most of the ineffectual computations on-the-fly, improving performance and energy efficiency compared to state-of-the-art high-performance accelerators [5]. The idea behind PRA is deceptively simple: use serial-parallel shift-and-add multiplication while skipping the zero bits of the serial input. However, a straightforward implementation based on shift-and-add multiplication yields unacceptable area, power and memory access overheads compared to a conventional bit-parallel design. PRA incorporates a set of design decisions to yield a practical, area and energy efficient design. Measurements demonstrate that for convolutional layers, PRA is 4.31


IEEE Computer Architecture Letters | 2017

Stripes: Bit-Serial Deep Neural Network Computing

Patrick Judd; Jorge Albericio; Andreas Moshovos

\times


international conference on embedded computer systems architectures modeling and simulation | 2014

Evaluating the memory system behavior of smartphone workloads

G. Narancic; Patrick Judd; D. Wu; Islam Atta; M. Elnacouzi; Jason Zebchuk; Jorge Albericio; N. Enright Jerger; Andreas Moshovos; K. Kutulakos; S. Gadelrab

faster than DaDianNao [5] (DaDN) using a 16-bit fixed-point representation. While PRA requires 1.68


design automation conference | 2018

Loom: exploiting weight and activation precisions to accelerate convolutional neural networks

Sayeh Sharify; Alberto Delmas Lascorz; Kevin Siu; Patrick Judd; Andreas Moshovos

\times


IEEE Micro | 2018

Value-Based Deep Learning Hardware Accelerators

Andreas Moshovos; Jorge Albericio; Patrick Judd; Alberto Delmas Lascorz; Sayeh Sharify; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger

more area than DaDN, the performance gains yield a 1.70


IEEE Computer | 2018

Exploiting Typical Values to Accelerate Deep Learning

Andreas Moshovos; Jorge Albericio; Patrick Judd; Alberto Delmas Lascorz; Sayeh Sharify; Zissis Poulos; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger

\times


arXiv: Learning | 2015

Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets

Patrick Judd; Jorge Albericio; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger; Raquel Urtasun; Andreas Moshovos

increase in energy efficiency in a 65nm technology. With 8-bit quantized activations, PRA is 2.25


international symposium on microarchitecture | 2016

Stripes: bit-serial deep neural network computing

Patrick Judd; Jorge Albericio; Tayler H. Hetherington; Tor M. Aamodt; Andreas Moshovos

\times


arXiv: Neural and Evolutionary Computing | 2017

Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks.

Alberto Delmas; Patrick Judd; Sayeh Sharify; Andreas Moshovos

faster and 1.31

Collaboration


Dive into the Patrick Judd's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Tayler H. Hetherington

University of British Columbia

View shared research outputs
Top Co-Authors

Avatar

Tor M. Aamodt

University of British Columbia

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge