Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Dustin Hillard is active.

Publication


Featured researches published by Dustin Hillard.


IEEE Transactions on Audio, Speech, and Language Processing | 2006

Enriching speech recognition with automatic detection of sentence boundaries and disfluencies

Yang Liu; Elizabeth Shriberg; Andreas Stolcke; Dustin Hillard; Mari Ostendorf; Mary P. Harper

Effective human and automatic processing of speech requires recovery of more than just the words. It also involves recovering phenomena such as sentence boundaries, filler words, and disfluencies, referred to as structural metadata. We describe a metadata detection system that combines information from different types of textual knowledge sources with information from a prosodic classifier. We investigate maximum entropy and conditional random field models, as well as the predominant hidden Markov model (HMM) approach, and find that discriminative models generally outperform generative models. We report system performance on both broadcast news and conversational telephone speech tasks, illustrating significant performance differences across tasks and as a function of recognizer performance. The results represent the state of the art, as assessed in the NIST RT-04F evaluation


north american chapter of the association for computational linguistics | 2003

Detection of agreement vs. disagreement in meetings: training with unlabeled data

Dustin Hillard; Mari Ostendorf; Elizabeth Shriberg

To support summarization of automatically transcribed meetings, we introduce a classifier to recognize agreement or disagreement utterances, utilizing both word-based and prosodic cues. We show that hand-labeling efforts can be minimized by using unsupervised training on a large unlabeled data set combined with supervised training on a small amount of data. For ASR transcripts with over 45% WER, the system recovers nearly 80% of agree/disagree utterances with a confusion rate of only 3%.


international conference on acoustics, speech, and signal processing | 2005

Structural metadata research in the EARS program

Yang Liu; Elizabeth Shriberg; Andreas Stolcke; Barbara Peskin; Jeremy Ang; Dustin Hillard; Mari Ostendorf; Marcus Tomalin; Philip C. Woodland; Mary P. Harper

Both human and automatic processing of speech require recognition of more than just words. In this paper we provide a brief overview of research on structural metadata extraction in the DARPA EARS rich transcription program. Tasks include detection of sentence boundaries, filler words, and disfluencies. Modeling approaches combine lexical, prosodic, and syntactic information, using various modeling techniques for knowledge source integration. The performance of these methods is evaluated by task, by data source (broadcast news versus spontaneous telephone conversations) and by whether transcriptions come from humans or from an (errorful) automatic speech recognizer. A representative sample of results shows that combining multiple knowledge sources (words, prosody, syntactic information) is helpful, that prosody is more helpful for news speech than for conversational speech, that word errors significantly impact performance, and that discriminative models generally provide benefit over maximum likelihood models. Important remaining issues, both technical and programmatic, are also discussed.


Journal of Information Technology & Politics | 2008

Computer-Assisted Topic Classification for Mixed-Methods Social Science Research

Dustin Hillard; Stephen Purpura; John Wilkerson

ABSTRACT Social scientists interested in mixed-methods research have traditionally turned to human annotators to classify the documents or events used in their analyses. The rapid growth of digitized government documents in recent years presents new opportunities for research but also new challenges. With more and more data coming online, relying on human annotators becomes prohibitively expensive for many tasks. For researchers interested in saving time and money while maintaining confidence in their results, we show how a particular supervised learning system can provide estimates of the class of each document (or event). This system maintains high classification accuracy and provides accurate estimates of document proportions, while achieving reliability levels associated with human efforts. We estimate that it lowers the costs of classifying large numbers of complex documents by 80% or more.


digital government research | 2006

Automated classification of congressional legislation

Stephen Purpura; Dustin Hillard

For social science researchers, content analysis and classification of United States Congressional legislative activities have been time consuming and costly. The Library of Congress THOMAS system provides detailed information about bills and laws, but its classification system, the Legislative Indexing Vocabulary (LIV), is geared toward information retrieval instead of the pattern or historical trend recognition that social scientists value. The same event (a bill) may be coded with many subjects at the same time, with little indication of its primary emphasis. In addition, because the LIV system has not been applied to other activities, it cannot be used to compare (for example) legislative issue attention to executive, media, or public issue attention.This paper presents the Congressional Bills Projects (www.congressionalbills.org) automated classification system. This system applies a topic spotting classification algorithm to the task of coding legislative activities into one of 226 subtopic areas. The algorithm uses a traditional bag-of-words document representation, an extensive set of human coded examples, and an exhaustive topic coding system developed for use by the Congressional Bills Project and the Policy Agendas Project (www.policyagendas.org). Experimental results demonstrate that the automated system is about as effective as human assessors, but with significant time and cost savings. The paper concludes by discussing challenges to moving the system into operational use.


international conference on acoustics, speech, and signal processing | 2008

Punctuating speech for information extraction

Benoit Favre; Ralph Grishman; Dustin Hillard; Heng Ji; Dilek Hakkani-Tür; Mari Ostendorf

This paper studies the effect of automatic sentence boundary detection and comma prediction on entity and relation extraction in speech. We show that punctuating the machine generated transcript according to maximum F-measure of period and comma annotation results in suboptimal information extraction. Precisely, period and comma decision thresholds can be chosen in order to improve the entity value score and the relation value score by 4% relative. Error analysis shows that preventing noun-phrase splitting by generating longer sentences and fewer commas can be harmful for IE performance. Indeed, it seems that missed punctuation allows syntactic parsers to merge noun-phrases and prevent the extraction of correct information.


north american chapter of the association for computational linguistics | 2007

iROVER: Improving System Combination with Classification

Dustin Hillard; Bjoern Hoffmeister; Mari Ostendorf; Ralf Schlueter; Hermann Ney

We present an improved system combination technique, iROVER, Our approach obtains significant improvements over ROVER, and is consistently better across varying numbers of component systems. A classifier is trained on features from the system lattices, and selects the final word hypothesis by learning cues to choose the system that is most likely to be correct at each word location. This approach achieves the best result published to date on the TC-STAR 2006 English speech recognition evaluation set.


international conference on acoustics, speech, and signal processing | 2007

Cross-Site and Intra-Site ASR System Combination: Comparisons on Lattice and 1-Best Methods

Björn Hoffmeister; Dustin Hillard; Stefan Hahn; Ralf Schlüter; Mari Ostendorf; Hermann Ney

We evaluate system combination techniques for automatic speech recognition using systems from multiple sites who participated in the TC-STAR 2006 evaluation. Both lattice and 1-best combination techniques are tested for cross-site and intra-site tasks. For pairwise combinations the lattice based approaches can outperform 1-best ROVER with confidence scores, but 1-best ROVER results are equal (or even better) when combining three or four systems.


north american chapter of the association for computational linguistics | 2004

Improving automatic sentence boundary detection with confusion networks

Dustin Hillard; Mari Ostendorf; Andreas Stolcke; Yang Liu; Elizabeth Shriberg

We extend existing methods for automatic sentence boundary detection by leveraging multiple recognizer hypotheses in order to provide robustness to speech recognition errors. For each hypothesized word sequence, an HMM is used to estimate the posterior probability of a sentence boundary at each word boundary. The hypotheses are combined using confusion networks to determine the overall most likely events. Experiments show improved detection of sentences for conversational telephone speech, though results are mixed for broadcast news.


spoken language technology workshop | 2006

IMPACT OF AUTOMATIC COMMA PREDICTION ON POS/NAME TAGGING OF SPEECH

Dustin Hillard; Zhongqiang Huang; Heng Ji; Ralph Grishman; Dilek Hakkani-Tür; Mary P. Harper; Mari Ostendorf; Wen Wang

This work looks at the impact of automatically predicted commas on part-of-speech (POS) and name tagging of speech recognition transcripts of Mandarin broadcast news. There is a significant gain in both POS and name tagging accuracy due to using automatically predicted commas over sentence boundary prediction alone. One difference between Mandarin and English is that there are two types of commas, and experiments here show that, while they can be reliably distinguished in automatic prediction, the distinction does not give a clear benefit for POS or name tagging.

Collaboration


Dive into the Dustin Hillard's collaboration.

Top Co-Authors

Avatar

Mari Ostendorf

University of Washington

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Elizabeth Shriberg

Ludwig Maximilian University of Munich

View shared research outputs
Top Co-Authors

Avatar

Yang Liu

University of Texas at Dallas

View shared research outputs
Top Co-Authors

Avatar

Heng Ji

Rensselaer Polytechnic Institute

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Barbara Peskin

University of California

View shared research outputs
Researchain Logo
Decentralizing Knowledge