George Tucker
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by George Tucker.
spoken language technology workshop | 2016
Ming Sun; Anirudh Raju; George Tucker; Sankaran Panchapagesan; Gengshen Fu; Arindam Mandal; Spyros Matsoukas; Nikko Strom; Shiv Vitaladevuni
We propose a max-pooling based loss function for training Long Short-Term Memory (LSTM) networks for small-footprint keyword spotting (KWS), with low CPU, memory, and latency requirements. The max-pooling loss training can be further guided by initializing with a cross-entropy loss trained network. A posterior smoothing based evaluation approach is employed to measure keyword spotting performance. Our experimental results show that LSTM models trained using cross-entropy loss or max-pooling loss outperform a cross-entropy loss trained baseline feed-forward Deep Neural Network (DNN). In addition, max-pooling loss trained LSTM with randomly initialized network performs better compared to cross-entropy loss trained LSTM. Finally, the max-pooling loss trained LSTM initialized with a cross-entropy pre-trained network shows the best performance, which yields 67:6% relative reduction compared to baseline feed-forward DNN in Area Under the Curve (AUC) measure.
conference of the international speech communication association | 2016
George Tucker; Minhua Wu; Ming Sun; Sankaran Panchapagesan; Gengshen Fu; Shiv Vitaladevuni
Several consumer speech devices feature voice interfaces that perform on-device keyword spotting to initiate user interactions. Accurate on-device keyword spotting within a tight CPU budget is crucial for such devices. Motivated by this, we investigated two ways to improve deep neural network (DNN) acoustic models for keyword spotting without increasing CPU usage. First, we used low-rank weight matrices throughout the DNN. This allowed us to increase representational power by increasing the number of hidden nodes per layer without changing the total number of multiplications. Second, we used knowledge distilled from an ensemble of much larger DNNs used only during training. We systematically evaluated these two approaches on a massive corpus of far-field utterances. Alone both techniques improve performance and together they combine to give significant reductions in false alarms and misses without increasing CPU or memory usage.
international conference on machine learning | 2018
George Tucker; Surya Bhupatiraju; Shixiang Gu; Richard E. Turner; Zoubin Ghahramani; Sergey Levine
Policy gradient methods are a widely used class of model-free reinforcement learning algorithms where a state-dependent baseline is used to reduce gradient estimator variance. Several recent papers extend the baseline to depend on both the state and action and suggest that this significantly reduces variance and improves sample efficiency without introducing bias into the gradient estimates. To better understand this development, we decompose the variance of the policy gradient estimator and numerically show that learned state-action-dependent baselines do not in fact reduce variance over a state-dependent baseline in commonly tested benchmark domains. We confirm this unexpected result by reviewing the open-source code accompanying these prior papers, and show that subtle implementation decisions cause deviations from the methods presented in the papers and explain the source of the previously observed empirical gains. Furthermore, the variance decomposition highlights areas for improvement, which we demonstrate by illustrating a simple change to the typical value function parameterization that can significantly improve performance.
arXiv: Neural and Evolutionary Computing | 2017
Gabriel Pereyra; George Tucker; Jan Chorowski; Lukasz Kaiser; Geoffrey E. Hinton
neural information processing systems | 2017
George Tucker; Andriy Mnih; Chris J. Maddison; John Lawson; Jascha Sohl-Dickstein
neural information processing systems | 2017
Chris J. Maddison; Dieterich Lawson; George Tucker; Nicolas Heess; Mohammad Norouzi; Andriy Mnih; Arnaud Doucet; Yee Whye Teh
international conference on acoustics, speech, and signal processing | 2018
Dieterich Lawson; George Tucker; Chung-Cheng Chiu; Colin Raffel; Kevin Swersky; Navdeep Jaitly
arXiv: Learning | 2017
Chris J. Maddison; Dieterich Lawson; George Tucker; Nicolas Heess; Arnaud Doucet; Andriy Minh; Yee Whye Teh
international conference on learning representations | 2018
Carlos Riquelme; George Tucker; Jasper Snoek
arXiv: Neural and Evolutionary Computing | 2018
Niru Maheswaranathan; Luke Metz; George Tucker; Jascha Sohl-Dickstein