Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Daniel Gillick is active.

Publication


Featured researches published by Daniel Gillick.


workshop on statistical machine translation | 2006

Why Generative Phrase Models Underperform Surface Heuristics

John DeNero; Daniel Gillick; James Zhang; Daniel Klein

We investigate why weights from generative models underperform heuristic estimates in phrase-based machine translation. We first propose a simple generative, phrase-based model and verify that its estimates are inferior to those given by surface statistics. The performance gap stems primarily from the addition of a hidden segmentation variable, which increases the capacity for overfitting during maximum likelihood training with EM. In particular, while word level models benefit greatly from re-estimation, phrase-level models do not: the crucial difference is that distinct word alignments cannot all be correct, while distinct segmentations can. Alternate segmentations rather than alternate alignments compete, resulting in increased deter-minization of the phrase table, decreased generalization, and decreased final BLEU score. We also show that interpolation of the two methods can result in a modest increase in BLEU score.


international conference on acoustics, speech, and signal processing | 2009

A global optimization framework for meeting summarization

Daniel Gillick; Korbinian Riedhammer; Benoit Favre; Dilek Hakkani-Tür

We introduce a model for extractive meeting summarization based on the hypothesis that utterances convey bits of information, or concepts. Using keyphrases as concepts weighted by frequency, and an integer linear program to determine the best set of utterances, that is, covering as many concepts as possible while satisfying a length constraint, we achieve ROUGE scores at least as good as a ROUGE-based oracle derived from human summaries. This brings us to a critical discussion of ROUGE and the future of extractive meeting summarization.


ieee automatic speech recognition and understanding workshop | 2011

Don't multiply lightly: Quantifying problems with the acoustic model assumptions in speech recognition

Daniel Gillick; Larry Gillick; Steven Wegmann

We describe a series of experiments simulating data from the standard Hidden Markov Model (HMM) framework used for speech recognition. Starting with a set of test transcriptions, we begin by simulating every step of the generative process. In each subsequent experiment, we substitute a real component for a simulated component (real state durations rather than simulating from the transition models, for example), and compare the word error rates of the resulting data, thus quantifying the relative costs of each modeling assumption. A novel sampling process allows us to test the independence assumptions of the HMM, which appear to present far more serious problems than the other data/model mismatches.


international conference on acoustics, speech, and signal processing | 2005

Speaker detection without models

Daniel Gillick; Stephen Stafford; Barbara Peskin

In order to capture sequential information and to take advantage of extended training data conditions, we developed an algorithm for speaker detection that scores a test segment by comparing it directly to similar instances of that speech in the training data. This non-parametric technique, though at an early stage in its development, achieves error rates close to 1% on the NIST 2001 extended data task and performs extremely well in combination with a standard Gaussian mixture model system. We also present a new scoring method that significantly improves performance by capturing only positive evidence.


ieee automatic speech recognition and understanding workshop | 2007

Integrating several annotation layers for statistical information distillation

Michael Levit; Dilek Hakkani-Tür; Gökhan Tür; Daniel Gillick

We present a sentence extraction algorithm for Information Distillation, a task where for a given templated query, relevant passages must be extracted from massive audio and textual document sources. For each sentence of the relevant documents (that are assumed to be known from the upstream stages) we employ statistical classification methods to estimate the extent of its relevance to the query, whereby two aspects of relevance are taken into account: the template (type) of the query and its slots (free-text descriptions of names, organizations, topic, events and so on, around which templates are centered). The idiosyncrasy of the presented method is in the choice of features used for classification. We extract our features from charts, compilations of elements from various annotation levels, such as word transcriptions, syntactic and semantic parses, and Information Extraction annotations. In our experiments we show that this integrated approach outperforms a purely lexical baseline by as much as 30% relative in terms of F-measure. We also investigate the algorithms behavior under noisy conditions, by comparing its performance on ASR output and on corresponding manual transcriptions.


international conference on acoustics, speech, and signal processing | 2012

Discriminative training for speech recognition is compensating for statistical dependence in the HMM framework

Daniel Gillick; Steven Wegmann; Larry Gillick

The parameters of the standard Hidden Markov Model framework for speech recognition are typically trained via Maximum Likelihood. However, better recognition performance is achievable with discriminative training criteria like Maximum Mutual Information or Minimum Phone Error. While it is generally accepted that these discriminative criteria are better suited to minimizing Word Error Rate, there is very little qualitative intuition for how the improvements are achieved. Through a series of “resampling” experiments, we show that discriminative training (MPE in particular) appears to be compensating for a specific incorrect assumption of the HMM-that speech frames are conditionally independent.


ieee automatic speech recognition and understanding workshop | 2005

ICSI'S 2005 speaker recognition system

Nikki Mirghafori; Andrew O. Hatch; Steven Stafford; Kofi Boakye; Daniel Gillick; Barbara Peskin

This paper describes ICSIs 2005 speaker recognition system, which was one of the top performing systems in the NIST 2005 speaker recognition evaluation. The system is a combination of four sub-systems: 1) a keyword conditional HMM system, 2) an SVM-based lattice phone n-gram system, 3) a sequential nonparametric system, and 4) a traditional cepstral GMM System, developed by SRI. The first three systems are designed to take advantage of higher-level and long-term information. We observe that their performance is significantly improved when there is more training data. In this paper, we describe these sub-systems and present results for each system alone and in combination on the speaker recognition evaluation (SRE) 2005 development and evaluation data sets


Computer Speech & Language | 2009

IXIR: A statistical information distillation system

Michael Levit; Dilek Hakkani-Tür; Gökhan Tür; Daniel Gillick

The task of information distillation is to extract snippets from massive multilingual audio and textual document sources that are relevant for a given templated query. We present an approach that focuses on the sentence extraction phase of the distillation process. It selects document sentences with respect to their relevance to a query via statistical classification with support vector machines. The distinguishing contribution of the approach is a novel method to generate classification features. The features are extracted from charts, compilations of elements from various annotation layers, such as word transcriptions, syntactic and semantic parses, and information extraction (IE) annotations. We describe a procedure for creating charts from documents and queries, while paying special attention to query slots (free-text descriptions of names, organizations, topic, events and so on, around which templates are centered), and suggest various types of classification features that can be extracted from these charts. While observing a 30% relative improvement due to non-lexical annotation layers, we perform a detailed analysis of the contributions of each of these layers to classification performance.


north american chapter of the association for computational linguistics | 2016

Exploring the Steps of Verb Phrase Ellipsis

Zhengzhong Liu; Edgar Gonzàlez Pellicer; Daniel Gillick

Verb Phrase Ellipsis is a well-studied topic in theoretical linguistics but has received little attention as a computational problem. Here we propose a decomposition of the overall resolution problem into three tasks—target detection, antecedent head resolution, and antecedent boundary detection—and implement a number of computational approaches for each one. We also explore the relationships among these tasks by attempting joint learning over different combinations. Our new decomposition of the problem yields significantly improved performance on publicly available datasets, including a newly contributed one.


asia information retrieval symposium | 2010

Re-ranking Summaries Based on Cross-Document Information Extraction

Heng Ji; Juan Liu; Benoit Favre; Daniel Gillick; Dilek Hakkani-Tür

This paper describes a novel approach of improving multi-document summarization based on cross-document information extraction (IE). We describe a method to automatically incorporate IE results into sentence ranking. Experiments have shown our integration methods can significantly improve a high-performing multi-document summarization system, according to the ROUGE-2 and ROUGE-SU4 metrics (7.38% relative improvement on ROUGE-2 recall), and the generated summaries are preferred by human subjects (0.78 higher TAC Content score and 0.11 higher Readability/Fluency score).

Collaboration


Dive into the Daniel Gillick's collaboration.

Top Co-Authors

Avatar

Benoit Favre

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Barbara Peskin

University of California

View shared research outputs
Top Co-Authors

Avatar

Daniel Klein

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

James Zhang

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Steven Wegmann

International Computer Science Institute

View shared research outputs
Researchain Logo
Decentralizing Knowledge