Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Stephen Tyree is active.

Publication


Featured researches published by Stephen Tyree.


international world wide web conferences | 2011

Parallel boosted regression trees for web search ranking

Stephen Tyree; Kilian Q. Weinberger; Kunal Agrawal; Jennifer Paykin

Gradient Boosted Regression Trees (GBRT) are the current state-of-the-art learning paradigm for machine learned web-search ranking - a domain notorious for very large data sets. In this paper, we propose a novel method for parallelizing the training of GBRT. Our technique parallelizes the construction of the individual regression trees and operates using the master-worker paradigm as follows. The data are partitioned among the workers. At each iteration, the worker summarizes its data-partition using histograms. The master processor uses these to build one layer of a regression tree, and then sends this layer to the workers, allowing the workers to build histograms for the next layer. Our algorithm carefully orchestrates overlap between communication and computation to achieve good performance. Since this approach is based on data partitioning, and requires a small amount of communication, it generalizes to distributed and shared memory machines, as well as clouds. We present experimental results on both shared memory machines and clusters for two large scale web search ranking data sets. We demonstrate that the loss in accuracy induced due to the histogram approximation in the regression tree creation can be compensated for through slightly deeper trees. As a result, we see no significant loss in accuracy on the Yahoo data sets and a very small reduction in accuracy for the Microsoft LETOR data. In addition, on shared memory machines, we obtain almost perfect linear speed-up with up to about 48 cores on the large data sets. On distributed memory machines, we get a speedup of 25 with 32 processors. Due to data partitioning our approach can scale to even larger data sets, on which one can reasonably expect even higher speedups.


international conference of the ieee engineering in medicine and biology society | 2013

Predicting a multi-parametric probability map of active tumor extent using random forests

Fred W. Prior; Sarah Jost Fouke; Tammie L.S. Benzinger; Alicia Boyd; Michael R. Chicoine; Sharath R. Cholleti; Matthew Kelsey; Bart Keogh; Lauren Kim; Mikhail Milchenko; David G. Politte; Stephen Tyree; Kilian Q. Weinberger; Daniel S. Marcus

Glioblastoma Mulitforme is highly infiltrative, making precise delineation of tumor margin difficult. Multimodality or multi-parametric MR imaging sequences promise an advantage over anatomic sequences such as post contrast enhancement as methods for determining the spatial extent of tumor involvement. In considering multi-parametric imaging sequences however, manual image segmentation and classification is time-consuming and prone to error. As a preliminary step toward integration of multi-parametric imaging into clinical assessments of primary brain tumors, we propose a machine-learning based multi-parametric approach that uses radiologist generated labels to train a classifier that is able to classify tissue on a voxel-wise basis and automatically generate a tumor segmentation. A random forests classifier was trained using a leave-one-out experimental paradigm. A simple linear classifier was also trained for comparison. The random forests classifier accurately predicted radiologist generated segmentations and tumor extent.


Archive | 2014

Approximation and Relaxation Approaches for Parallel and Distributed Machine Learning

Stephen Tyree

OF THE DISSERTATION Approximation and Relaxation Approaches for Parallel and Distributed Machine Learning by Stephen W. Tyree Doctor of Philosophy in Computer Science Washington University in St. Louis, 2014 Professor Kilian Q. Weinberger, Chair Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to choose simpler, less powerful models, e.g. linear models, in order to process more training examples in a limited time. In this work, we introduce parallelism to the training of non-linear models by leveraging a different tradeoff—approximation. We demonstrate various techniques by which non-linear models can be made amenable to larger data sets and significantly more training parallelism by strategically introducing approximation in certain optimization steps. For gradient boosted regression tree ensembles, we replace precise selection of tree splits with a coarse-grained, approximate split selection, yielding both faster sequential training and a significant increase in parallelism, in the distributed setting in particular. For metric learning with nearest neighbor classification, rather than explicitly train a neighborhood structure we leverage the implicit neighborhood structure induced by task-specific random forest classifiers, yielding a highly parallel method for metric learning. For support vector machines, we follow existing work to learn a reduced basis set with extremely high parallelism, particularly on GPUs, via existing linear algebra libraries.


neural information processing systems | 2012

Non-linear Metric Learning

Dor Kedem; Stephen Tyree; Fei Sha; Gert R. G. Lanckriet; Kilian Q. Weinberger


international conference on machine learning | 2013

Learning with Marginalized Corrupted Features

Laurens van der Maaten; Minmin Chen; Stephen Tyree; Kilian Q. Weinberger


arXiv: Learning | 2015

Compressing Convolutional Neural Networks.

Wenlin Chen; James T. Wilson; Stephen Tyree; Kilian Q. Weinberger; Yixin Chen


international conference on machine learning | 2014

Stochastic Neighbor Compression

Matt J. Kusner; Stephen Tyree; Kilian Q. Weinberger; Kunal Agrawal


arXiv: Learning | 2016

Pruning Convolutional Neural Networks for Resource Efficient Transfer Learning.

Pavlo Molchanov; Stephen Tyree; Tero Karras; Timo Aila; Jan Kautz


arXiv: Learning | 2014

Parallel Support Vector Machines in Practice

Stephen Tyree; Jacob R. Gardner; Kilian Q. Weinberger; Kunal Agrawal; John Tran


arXiv: Learning | 2016

GA3C: GPU-based A3C for Deep Reinforcement Learning

Mohammad Babaeizadeh; Iuri Frosio; Stephen Tyree; Jason Clemons; Jan Kautz

Collaboration


Dive into the Stephen Tyree's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kunal Agrawal

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Matt J. Kusner

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Jan Kautz

University College London

View shared research outputs
Top Co-Authors

Avatar

Jacob R. Gardner

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Minmin Chen

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Nicholas I. Kolkin

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Laurens van der Maaten

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Alicia Boyd

Washington University in St. Louis

View shared research outputs
Researchain Logo
Decentralizing Knowledge