Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Zhixiang Eddie Xu is active.

Publication


Featured researches published by Zhixiang Eddie Xu.


knowledge discovery and data mining | 2014

Gradient boosted feature selection

Zhixiang Eddie Xu; Gao Huang; Kilian Q. Weinberger; Alice X. Zheng

A feature selection algorithm should ideally satisfy four conditions: reliably extract relevant features; be able to identify non-linear feature interactions; scale linearly with the number of features and dimensions; allow the incorporation of known sparsity structure. In this work we propose a novel feature selection algorithm, Gradient Boosted Feature Selection (GBFS), which satisfies all four of these requirements. The algorithm is flexible, scalable, and surprisingly straight-forward to implement as it is based on a modification of Gradient Boosted Trees. We evaluate GBFS on several real world data sets and show that it matches or outperforms other state of the art feature selection algorithms. Yet it scales to larger data set sizes and naturally allows for domain-specific side information.


conference on information and knowledge management | 2012

From sBoW to dCoT marginalized encoders for text representation

Zhixiang Eddie Xu; Minmin Chen; Kilian Q. Weinberger; Fei Sha

In text mining, information retrieval, and machine learning, text documents are commonly represented through variants of sparse Bag of Words (sBoW) vectors (e.g. TF-IDF [1]). Although simple and intuitive, sBoW style representations suffer from their inherent over-sparsity and fail to capture word-level synonymy and polysemy. Especially when labeled data is limited (e.g. in document classification), or the text documents are short (e.g. emails or abstracts), many features are rarely observed within the training corpus. This leads to overfitting and reduced generalization accuracy. In this paper we propose Dense Cohort of Terms (dCoT), an unsupervised algorithm to learn improved sBoW document features. dCoT explicitly models absent words by removing and reconstructing random sub-sets of words in the unlabeled corpus. With this approach, dCoT learns to reconstruct frequent words from co-occurring infrequent words and maps the high dimensional sparse sBoW vectors into a low-dimensional dense representation. We show that the feature removal can be marginalized out and that the reconstruction can be solved for in closed-form. We demonstrate empirically, on several benchmark datasets, that dCoT features significantly improve the classification accuracy across several document classification tasks.


european conference on machine learning | 2014

Transductive Minimax Probability Machine

Gao Huang; Shiji Song; Zhixiang Eddie Xu; Kilian Q. Weinberger

The Minimax Probability Machine (MPM) is an elegant machine learning algorithm for inductive learning. It learns a classifier that minimizes an upper bound on its own generalization error. In this paper, we extend its celebrated inductive formulation to an equally elegant transductive learning algorithm. In the transductive setting, the label assignment of a test set is already optimized during training. This optimization problem is an intractable mixed-integer programming. Thus, we provide an efficient label-switching approach to solve it approximately. The resulting method scales naturally to large data sets and is very efficient to run. In comparison with nine competitive algorithms on eleven data sets, we show that the proposed Transductive MPM (TMPM) almost outperforms all the other algorithms in both accuracy and speed.


Archive | 2014

Supervised Machine Learning Under Test-Time Resource Constraints: A Trade-off Between Accuracy and Cost

Zhixiang Eddie Xu

OF THE DISSERTATION Supervised Machine Learning Under Test-Time Resource Constraints: A Trade-off Between Accuracy and Cost by Zhixiang (Eddie) Xu Doctor of Philosophy in Computer Science Washington University in St. Louis, 2014 Research Advisor: Professor Kilian Q. Weinberger, Chair The past decade has witnessed how the field of machine learning has established itself as a necessary component in several multi-billion-dollar industries. The real-world industrial setting introduces an interesting new problem to machine learning research: computational resources must be budgeted and cost must be strictly accounted for during test-time. A typical problem is that if an application consumes x additional units of cost during test-time, but will improve accuracy by y percent, should the additional x resources be allocated? The core of this problem is a trade-off between accuracy and cost. In this thesis, we examine components of test-time cost, and develop different strategies to manage this trade-off. We first investigate test-time cost and discover that it typically consists of two parts: feature extraction cost and classifier evaluation cost. The former reflects the computational efforts of transforming data instances to feature vectors, and could be highly variable when features are heterogeneous. The latter reflects the effort of evaluating a classifier, which could be substantial, in particular nonparametric algorithms. We then propose three strategies ix to explicitly trade-off accuracy and the two components of test-time cost during classifier training. To budget the feature extraction cost, we first introduce two algorithms: GreedyMiser [132] and Anytime Representation Learning (AFR)[135]. GreedyMiser employs a strategy that incorporates the extraction cost information during classifier training to explicitly minimize the test-time cost. AFR extends GreedyMiser to learn a cost-sensitive feature representation rather than a classifier, and turns traditional Support Vector Machines (SVM) [110] into testtime cost-sensitive anytime classifiers. GreedyMiser and AFR are evaluated on two real-world data sets from two different application domains, and both achieve record performance. We then introduce Cost Sensitive Tree of Classifiers (CSTC)[134] and Cost Sensitive Cascade of Classifiers (CSCC)[137], which share a common strategy that trades-off the accuracy and the amortized test-time cost. CSTC introduces a tree structure and directs test inputs along different tree traversal paths, each is optimized for a specific sub-partition of the input space, extracting different, specialized subsets of features. CSCC extends CSTC and builds a linear cascade, instead of a tree, to cope with class-imbalanced binary classification tasks. Since both CSTC and CSCC extract different features for different inputs, the amortized test-time cost is greatly reduced while maintaining high accuracy. Both approaches out-perform the current state-of-the-art on real-world data sets. To trade-off accuracy and high classifier evaluation cost of nonparametric classifiers, we propose a model compression strategy and develop Compressed Vector Machines (CVM). CVM focuses on the nonparametric kernel Support Vector Machines (SVM), whose testtime evaluation cost is typically substantial when learned from large training sets. CVM is a post-processing algorithm which compresses the learned SVM model by reducing and


international conference on machine learning | 2013

Cost-Sensitive Tree of Classifiers

Zhixiang Eddie Xu; Matt J. Kusner; Kilian Q. Weinberger; Minmin Chen


Journal of Machine Learning Research | 2014

Classifier cascades and trees for minimizing feature evaluation cost

Zhixiang Eddie Xu; Matt J. Kusner; Kilian Q. Weinberger; Minmin Chen; Olivier Chapelle


arXiv: Machine Learning | 2012

Distance Metric Learning for Kernel Machines

Zhixiang Eddie Xu; Kilian Q. Weinberger; Olivier Chapelle


national conference on artificial intelligence | 2014

Feature-cost sensitive learning with submodular trees of classifiers

Matt J. Kusner; Wenlin Chen; Quan Zhou; Zhixiang Eddie Xu; Kilian Q. Weinberger; Yixin Chen


international conference on machine learning | 2013

Anytime Representation Learning

Zhixiang Eddie Xu; Matt J. Kusner; Gao Huang; Kilian Q. Weinberger


Journal of Machine Learning Research | 2015

Marginalizing stacked linear denoising autoencoders

Minmin Chen; Kilian Q. Weinberger; Zhixiang Eddie Xu; Fei Sha

Collaboration


Dive into the Zhixiang Eddie Xu's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Minmin Chen

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Fei Sha

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Matt J. Kusner

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jacob R. Gardner

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Stephen Tyree

Washington University in St. Louis

View shared research outputs
Top Co-Authors

Avatar

Wenlin Chen

Washington University in St. Louis

View shared research outputs
Researchain Logo
Decentralizing Knowledge