Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where John M. Conroy is active.

Publication


Featured researches published by John M. Conroy.


international acm sigir conference on research and development in information retrieval | 2001

Text summarization via hidden Markov models

John M. Conroy; Dianne P. O'Leary

A sentence extract summary of a document is a subset of the documents sentences that contains the main ideas in the document. We present an approach to generating such summaries, a hidden Markov model that judges the likelihood that each sentence should be contained in the summary. We compare the results of this method with summaries generated by humans, showing that we obtain significantly higher agreement than do earlier methods.


international conference on computational linguistics | 2008

Mind the Gap: Dangers of Divorcing Evaluations of Summary Content from Linguistic Quality

John M. Conroy; Hoa Trang Dang

In this paper, we analyze the state of current human and automatic evaluation of topic-focused summarization in the Document Understanding Conference main task for 2005--2007. The analyses show that while ROUGE has very strong correlation with responsiveness for both human and automatic summaries, there is a significant gap in responsiveness between humans and systems which is not accounted for by the ROUGE metrics. In addition to teasing out gaps in the current automatic evaluation, we propose a method to maximize the strength of current automatic evaluations by using the method of canonical correlation. We apply this new evaluation method, which we call ROSE (ROUGE Optimal Summarization Evaluation), to find the optimal linear combination of ROUGE scores to maximize correlation with human responsiveness.


international conference on computational linguistics | 2008

Arabic/English multi-document summarization with CLASSY: the past and the future

Judith D. Schlesinger; Dianne P. O'Leary; John M. Conroy

Automatic document summarization has become increasingly important due to the quantity of written material generated worldwide. Generating good quality summaries enables users to cope with larger amounts of information. English-document summarization is a difficult task. Yet it is not sufficient. Environmental, economic, and other global issues make it imperative for English speakers to understand how other countries and cultures perceive and react to important events. CLASSY (Clustering, Linguistics, And Statistics for Summarization Yield) is an automatic, extract-generating, summarization system that uses linguistic trimming and statistical methods to generate generic or topic(/query)-driven summaries for single documents or clusters of documents. CLASSY has performed well in the Document Understanding Conference (DUC) evaluations and the Multi-lingual (Arabic/English) Summarization Evaluations (MSE). We present a description of CLASSY. We follow this with experiments and results from the MSE evaluations and conclude with a discussion of on-going work to improve the quality of the summaries-both Englishonly and multi-lingual-that CLASSY generates.


natural language generation | 2007

Measuring Variability in Sentence Ordering for News Summarization

Nitin Madnani; Rebecca J. Passonneau; Necip Fazil Ayan; John M. Conroy; Bonnie J. Dorr; Judith L. Klavans; Dianne P. O'Leary; Judith D. Schlesinger

The issue of sentence ordering is an important one for natural language tasks such as multi-document summarization, yet there has not been a quantitative exploration of the range of acceptable sentence orderings for short texts. We present results of a sentence reordering experiment with three experimental conditions. Our findings indicate a very high degree of variability in the orderings that the eighteen subjects produce. In addition, the variability of reorderings is significantly greater when the initial ordering seen by subjects is different from the original summary. We conclude that evaluation of sentence ordering should use multiple reference orderings. Our evaluation presents several metrics that might prove useful in assessing against multiple references. We conclude with a deeper set of questions: (a) what sorts of independent assessments of quality of the different reference orderings could be made and (b) whether a large enough test set would obviate the need for such independent means of quality assessment.


international conference on data mining | 2012

OCCAMS -- An Optimal Combinatorial Covering Algorithm for Multi-document Summarization

Sashka T. Davis; John M. Conroy; Judith D. Schlesinger

OCCAMS is a new algorithm for the Multi-Document Summarization (MDS) problem. We use Latent Semantic Analysis (LSA) to produce term weights which identify the main theme(s) of a set of documents. These are used by our heuristic for extractive sentence selection which borrows techniques from combinatorial optimization to select a set of sentences such that the combined weight of the terms covered is maximized while redundancy is minimized. OCCAMS outperforms CLASSY11 on DUC/TAC data for nearly all years since 2005, where CLASSY11 is the best human-rated system of TAC 2011. OCCAMS also delivers higher ROUGE scores than all human-generated summaries for TAC 2011. We show that if the combinatorial component of OCCAMS, which computes the extractive summary, is given true weights of terms, then the quality of the summaries generated outperforms all human generated summaries for all years using ROUGE-2, ROUGE-SU4, and a coverage metric. We introduce this new metric based on term coverage and demonstrate that a simple bi-gram instantiation achieves a statistically significant higher Pearson correlation with overall responsiveness than ROUGE on the TAC data.


IEEE Intelligent Systems | 2003

Machine and human performance for single and multidocument summarization

Judith D. Schlesinger; John M. Conroy; Mary Ellen Okurowski; Dianne P. O'Leary

The DUC 2002 evaluation revealed numerous language-processing challenges that impact text summarization. The authors examine the techniques used in a multidocument summarization system they developed and its performance at DUC 2002. They also discuss the need for regularization of human summaries.


Computational Linguistics | 2011

Nouveau-rouge: A novelty metric for update summarization

John M. Conroy; Judith D. Schlesinger; Dianne P. O'Leary

An update summary should provide a fluent summarization of new information on a time-evolving topic, assuming that the reader has already reviewed older documents or summaries. In 2007 and 2008, an annual summarization evaluation included an update summarization task. Several participating systems produced update summaries indistinguishable from human-generated summaries when measured using ROUGE. However, no machine system performed near human-level performance in manual evaluations such as pyramid and overall responsiveness scoring. We present a metric called Nouveau-ROUGE that improves correlation with manual evaluation metrics and can be used to predict both the pyramid score and overall responsiveness for update summaries. Nouveau-ROUGE can serve as a less expensive surrogate for manual evaluations when comparing existing systems and when developing new ones.


Laboratory Investigation | 2000

Chromosome Identification Using Hidden Markov Models: Comparison with Neural Networks, Singular Value Decomposition, Principal Components Analysis, and Fisher Discriminant Analysis

John M. Conroy; Tamara G. Kolda; Dianne P. O'Leary; Timothy J. O'Leary

The analysis of G-banded chromosomes remains the most important tool available to the clinical cytogeneticist. The analysis is laborious when performed manually, and the utility of automated chromosome identification algorithms has been limited by the fact that classification accuracy of these methods seldom exceeds about 80% in routine practice. In this study, we use four new approaches to automated chromosome identification — singular value decomposition (SVD), principal components analysis (PCA), Fisher discriminant analysis (FDA), and hidden Markov models (HMM) — to classify three well-known chromosome data sets (Philadelphia, Edinburgh, and Copenhagen), comparing these approaches with the use of neural networks (NN). We show that the HMM is a particularly robust approach to identification that attains classification accuracies of up to 97% for normal chromosomes and retains classification accuracies of up to 95% when chromosome telomeres are truncated or small portions of the chromosome are inverted. This represents a substantial improvement of the classification accuracy for normal chromosomes, and a doubling in classification accuracy for truncated chromosomes and those with inversions, as compared with NN-based methods. HMMs thus appear to be a promising approach for the automated identification of both normal and abnormal G-banded chromosomes.


north american chapter of the association for computational linguistics | 2003

QCS: a tool for querying, clustering, and summarizing documents

Daniel M. Dunlavy; John M. Conroy; Dianne P. O'Leary

The QCS information retrieval (IR) system is presented as a tool for querying, clustering, and summarizing document sets. QCS has been developed as a modular development framework, and thus facilitates the inclusion of new technologies targeting these three IR tasks. Details of the system architecture, the QCS interface, and preliminary results are presented.


Algorithms | 2012

Better Metrics to Automatically Predict the Quality of a Text Summary

Peter A. Rankel; John M. Conroy; Judith D. Schlesinger

In this paper we demonstrate a family of metrics for estimating the quality of a text summary relative to one or more human-generated summaries. The improved metrics are based on features automatically computed from the summaries to measure content and linguistic quality. The features are combined using one of three methods—robust regression, non-negative least squares, or canonical correlation, an eigenvalue method. The new metrics significantly outperform the previous standard for automatic text summarization evaluation, ROUGE.

Collaboration


Dive into the John M. Conroy's collaboration.

Top Co-Authors

Avatar

Daniel M. Dunlavy

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

Hoa Trang Dang

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Tamara G. Kolda

Sandia National Laboratories

View shared research outputs
Top Co-Authors

Avatar

Timothy J. O'Leary

Veterans Health Administration

View shared research outputs
Top Co-Authors

Avatar

Ani Nenkova

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Karolina Owczarzak

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Yi-Kai Liu

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge