Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Gregory Druck is active.

Publication


Featured researches published by Gregory Druck.


international acm sigir conference on research and development in information retrieval | 2008

Learning from labeled features using generalized expectation criteria

Gregory Druck; Gideon S. Mann; Andrew McCallum

It is difficult to apply machine learning to new domains because often we lack labeled problem instances. In this paper, we provide a solution to this problem that leverages domain knowledge in the form of affinities between input features and classes. For example, in a baseball vs. hockey text classification problem, even without any labeled data, we know that the presence of the word puck is a strong indicator of hockey. We refer to this type of domain knowledge as a labeled feature. In this paper, we propose a method for training discriminative probabilistic models with labeled features and unlabeled instances. Unlike previous approaches that use labeled features to create labeled pseudo-instances, we use labeled features directly to constrain the models predictions on unlabeled instances. We express these soft constraints using generalized expectation (GE) criteria --- terms in a parameter estimation objective function that express preferences on values of a model expectation. In this paper we train multinomial logistic regression models using GE criteria, but the method we develop is applicable to other discriminative probabilistic models. The complete objective function also includes a Gaussian prior on parameters, which encourages generalization by spreading parameter weight to unlabeled features. Experimental results on text classification data sets show that this method outperforms heuristic approaches to training classifiers with labeled features. Experiments with human annotators show that it is more beneficial to spend limited annotation time labeling features rather than labeling instances. For example, after only one minute of labeling features, we can achieve 80% accuracy on the ibm vs. mac text classification problem using GE-FL, whereas ten minutes labeling documents results in an accuracy of only 77%


knowledge discovery and data mining | 2007

Semi-supervised classification with hybrid generative/discriminative methods

Gregory Druck; Chris Pal; Andrew McCallum; Xiaojin Zhu

We compare two recently proposed frameworks for combining generative and discriminative probabilistic classifiers and apply them to semi-supervised classification. In both cases we explore the tradeoff between maximizing a discriminative likelihood of labeled data and a generative likelihood of labeled and unlabeled data. While prominent semi-supervised learning methods assume low density regions between classes or are subject to generative modeling assumptions, we conjecture that hybrid generative/discriminative methods allow semi-supervised learning in the presence of strongly overlapping classes and reduce the risk of modeling structure in the unlabeled data that is irrelevant for the specific classification task of interest. We apply both hybrid approaches within naively structured Markov random field models and provide a thorough empirical comparison with two well-known semi-supervised learning methods on six text classification tasks. A semi-supervised hybrid generative/discriminative method provides the best accuracy in 75% of the experiments, and the multi-conditional learning hybrid approach achieves the highest overall mean accuracy across all tasks.


conference on information and knowledge management | 2011

Toward interactive training and evaluation

Gregory Druck; Andrew McCallum

Machine learning often relies on costly labeled data, and this impedes its application to new classification and information extraction problems. This has motivated the development of methods for leveraging abundant prior knowledge about these problems, including methods for lightly supervised learning using model expectation constraints. Building on this work, we envision an interactive training paradigm in which practitioners perform evaluation, analyze errors, and provide and refine expectation constraints in a closed loop. In this paper, we focus on several key subproblems in this paradigm that can be cast as selecting a representative sample of the unlabeled data for the practitioner to inspect. To address these problems, we propose stratified sampling methods that use model expectations as a proxy for latent output variables. In classification and sequence labeling experiments, these sampling strategies reduce accuracy evaluation effort by as much as 53%, provide more reliable estimates of


empirical methods in natural language processing | 2009

Active Learning by Labeling Features

Gregory Druck; Burr Settles; Andrew McCallum

F_1


national conference on artificial intelligence | 2006

Multi-conditional learning: generative/discriminative training for clustering and classification

Andrew McCallum; Chris Pal; Gregory Druck; Xuerui Wang

for rare labels, and aid in the specification and refinement of constraints.


uncertainty in artificial intelligence | 2009

Alternating projections for learning with expectation constraints

Kedar Bellare; Gregory Druck; Andrew McCallum


international joint conference on natural language processing | 2009

Semi-supervised Learning of Dependency Parsers using Generalized Expectation Criteria

Gregory Druck; Gideon S. Mann; Andrew McCallum


Archive | 2007

Generalized Expectation Criteria

Andrew McCallum; Gideon S. Mann; Gregory Druck


Archive | 2008

Learning to Predict the Quality of Contributions to Wikipedia

Gregory Druck; Gerome Miklau; Andrew McCallum


international conference on machine learning | 2010

High-Performance Semi-Supervised Learning using Discriminatively Constrained Generative Models

Gregory Druck; Andrew McCallum

Collaboration


Dive into the Gregory Druck's collaboration.

Top Co-Authors

Avatar

Andrew McCallum

University of Massachusetts Amherst

View shared research outputs
Top Co-Authors

Avatar

Chris Pal

École Polytechnique de Montréal

View shared research outputs
Top Co-Authors

Avatar

Burr Settles

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Gerome Miklau

University of Massachusetts Amherst

View shared research outputs
Top Co-Authors

Avatar

Kedar Bellare

University of Massachusetts Amherst

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xiaojin Zhu

University of Wisconsin-Madison

View shared research outputs
Researchain Logo
Decentralizing Knowledge