Stephen Tyree
Washington University in St. Louis
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Stephen Tyree.
international world wide web conferences | 2011
Stephen Tyree; Kilian Q. Weinberger; Kunal Agrawal; Jennifer Paykin
Gradient Boosted Regression Trees (GBRT) are the current state-of-the-art learning paradigm for machine learned web-search ranking - a domain notorious for very large data sets. In this paper, we propose a novel method for parallelizing the training of GBRT. Our technique parallelizes the construction of the individual regression trees and operates using the master-worker paradigm as follows. The data are partitioned among the workers. At each iteration, the worker summarizes its data-partition using histograms. The master processor uses these to build one layer of a regression tree, and then sends this layer to the workers, allowing the workers to build histograms for the next layer. Our algorithm carefully orchestrates overlap between communication and computation to achieve good performance. Since this approach is based on data partitioning, and requires a small amount of communication, it generalizes to distributed and shared memory machines, as well as clouds. We present experimental results on both shared memory machines and clusters for two large scale web search ranking data sets. We demonstrate that the loss in accuracy induced due to the histogram approximation in the regression tree creation can be compensated for through slightly deeper trees. As a result, we see no significant loss in accuracy on the Yahoo data sets and a very small reduction in accuracy for the Microsoft LETOR data. In addition, on shared memory machines, we obtain almost perfect linear speed-up with up to about 48 cores on the large data sets. On distributed memory machines, we get a speedup of 25 with 32 processors. Due to data partitioning our approach can scale to even larger data sets, on which one can reasonably expect even higher speedups.
international conference of the ieee engineering in medicine and biology society | 2013
Fred W. Prior; Sarah Jost Fouke; Tammie L.S. Benzinger; Alicia Boyd; Michael R. Chicoine; Sharath R. Cholleti; Matthew Kelsey; Bart Keogh; Lauren Kim; Mikhail Milchenko; David G. Politte; Stephen Tyree; Kilian Q. Weinberger; Daniel S. Marcus
Glioblastoma Mulitforme is highly infiltrative, making precise delineation of tumor margin difficult. Multimodality or multi-parametric MR imaging sequences promise an advantage over anatomic sequences such as post contrast enhancement as methods for determining the spatial extent of tumor involvement. In considering multi-parametric imaging sequences however, manual image segmentation and classification is time-consuming and prone to error. As a preliminary step toward integration of multi-parametric imaging into clinical assessments of primary brain tumors, we propose a machine-learning based multi-parametric approach that uses radiologist generated labels to train a classifier that is able to classify tissue on a voxel-wise basis and automatically generate a tumor segmentation. A random forests classifier was trained using a leave-one-out experimental paradigm. A simple linear classifier was also trained for comparison. The random forests classifier accurately predicted radiologist generated segmentations and tumor extent.
Archive | 2014
Stephen Tyree
OF THE DISSERTATION Approximation and Relaxation Approaches for Parallel and Distributed Machine Learning by Stephen W. Tyree Doctor of Philosophy in Computer Science Washington University in St. Louis, 2014 Professor Kilian Q. Weinberger, Chair Large scale machine learning requires tradeoffs. Commonly this tradeoff has led practitioners to choose simpler, less powerful models, e.g. linear models, in order to process more training examples in a limited time. In this work, we introduce parallelism to the training of non-linear models by leveraging a different tradeoff—approximation. We demonstrate various techniques by which non-linear models can be made amenable to larger data sets and significantly more training parallelism by strategically introducing approximation in certain optimization steps. For gradient boosted regression tree ensembles, we replace precise selection of tree splits with a coarse-grained, approximate split selection, yielding both faster sequential training and a significant increase in parallelism, in the distributed setting in particular. For metric learning with nearest neighbor classification, rather than explicitly train a neighborhood structure we leverage the implicit neighborhood structure induced by task-specific random forest classifiers, yielding a highly parallel method for metric learning. For support vector machines, we follow existing work to learn a reduced basis set with extremely high parallelism, particularly on GPUs, via existing linear algebra libraries.
neural information processing systems | 2012
Dor Kedem; Stephen Tyree; Fei Sha; Gert R. G. Lanckriet; Kilian Q. Weinberger
international conference on machine learning | 2013
Laurens van der Maaten; Minmin Chen; Stephen Tyree; Kilian Q. Weinberger
arXiv: Learning | 2015
Wenlin Chen; James T. Wilson; Stephen Tyree; Kilian Q. Weinberger; Yixin Chen
international conference on machine learning | 2014
Matt J. Kusner; Stephen Tyree; Kilian Q. Weinberger; Kunal Agrawal
arXiv: Learning | 2016
Pavlo Molchanov; Stephen Tyree; Tero Karras; Timo Aila; Jan Kautz
arXiv: Learning | 2014
Stephen Tyree; Jacob R. Gardner; Kilian Q. Weinberger; Kunal Agrawal; John Tran
arXiv: Learning | 2016
Mohammad Babaeizadeh; Iuri Frosio; Stephen Tyree; Jason Clemons; Jan Kautz