Patrick Judd
University of Toronto
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Patrick Judd.
international symposium on computer architecture | 2016
Jorge Albericio; Patrick Judd; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger; Andreas Moshovos
This work observes that a large fraction of the computations performed by Deep Neural Networks (DNNs) are intrinsically ineffectual as they involve a multiplication where one of the inputs is zero. This observation motivates Cnvolutin (CNV), a value-based approach to hardware acceleration that eliminates most of these ineffectual operations, improving performance and energy over a state-of-the-art accelerator with no accuracy loss. CNV uses hierarchical data-parallel units, allowing groups of lanes to proceed mostly independently enabling them to skip over the ineffectual computations. A co-designed data storage format encodes the computation elimination decisions taking them off the critical path while avoiding control divergence in the data parallel units. Combined, the units and the data storage format result in a data-parallel architecture that maintains wide, aligned accesses to its memory hierarchy and that keeps its data lanes busy. By loosening the ineffectual computation identification criterion, CNV enables further performance and energy efficiency improvements, and more so if a loss in accuracy is acceptable. Experimental measurements over a set of state-of-the-art DNNs for image classification show that CNV improves performance over a state-of-the-art accelerator from 1.24× to 1.55× and by 1.37× on average without any loss in accuracy by removing zero-valued operand multiplications alone. While CNV incurs an area overhead of 4.49%, it improves overall EDP (Energy Delay Product) and ED2P (Energy Delay Squared Product) on average by 1.47× and 2.01×, respectively. The average performance improvements increase to 1.52× without any loss in accuracy with a broader ineffectual identification policy. Further improvements are demonstrated with a loss in accuracy.
international symposium on microarchitecture | 2017
Jorge Albericio; Alberto Delmas; Patrick Judd; Sayeh Sharify; Gerard O'Leary; Roman Genov; Andreas Moshovos
Deep Neural Networks expose a high degree of parallelism, making them amenable to highly data parallel architectures. However, data-parallel architectures often accept inefficiency in individual computations for the sake of overall efficiency. We show that on average, activation values of convolutional layers during inference in modern Deep Convolutional Neural Networks (CNNs) contain 92% zero bits. Processing these zero bits entails ineffectual computations that could be skipped. We propose Pragmatic (PRA), a massively data-parallel architecture that eliminates most of the ineffectual computations on-the-fly, improving performance and energy efficiency compared to state-of-the-art high-performance accelerators [5]. The idea behind PRA is deceptively simple: use serial-parallel shift-and-add multiplication while skipping the zero bits of the serial input. However, a straightforward implementation based on shift-and-add multiplication yields unacceptable area, power and memory access overheads compared to a conventional bit-parallel design. PRA incorporates a set of design decisions to yield a practical, area and energy efficient design. Measurements demonstrate that for convolutional layers, PRA is 4.31
IEEE Computer Architecture Letters | 2017
Patrick Judd; Jorge Albericio; Andreas Moshovos
\times
international conference on embedded computer systems architectures modeling and simulation | 2014
G. Narancic; Patrick Judd; D. Wu; Islam Atta; M. Elnacouzi; Jason Zebchuk; Jorge Albericio; N. Enright Jerger; Andreas Moshovos; K. Kutulakos; S. Gadelrab
faster than DaDianNao [5] (DaDN) using a 16-bit fixed-point representation. While PRA requires 1.68
design automation conference | 2018
Sayeh Sharify; Alberto Delmas Lascorz; Kevin Siu; Patrick Judd; Andreas Moshovos
\times
IEEE Micro | 2018
Andreas Moshovos; Jorge Albericio; Patrick Judd; Alberto Delmas Lascorz; Sayeh Sharify; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger
more area than DaDN, the performance gains yield a 1.70
IEEE Computer | 2018
Andreas Moshovos; Jorge Albericio; Patrick Judd; Alberto Delmas Lascorz; Sayeh Sharify; Zissis Poulos; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger
\times
arXiv: Learning | 2015
Patrick Judd; Jorge Albericio; Tayler H. Hetherington; Tor M. Aamodt; Natalie D. Enright Jerger; Raquel Urtasun; Andreas Moshovos
increase in energy efficiency in a 65nm technology. With 8-bit quantized activations, PRA is 2.25
international symposium on microarchitecture | 2016
Patrick Judd; Jorge Albericio; Tayler H. Hetherington; Tor M. Aamodt; Andreas Moshovos
\times
arXiv: Neural and Evolutionary Computing | 2017
Alberto Delmas; Patrick Judd; Sayeh Sharify; Andreas Moshovos
faster and 1.31