Gregory Marton
Massachusetts Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Gregory Marton.
international acm sigir conference on research and development in information retrieval | 2003
Stefanie Tellex; Boris Katz; Jimmy J. Lin; Aaron Fernandes; Gregory Marton
Passage retrieval is an important component common to many question answering systems. Because most evaluations of question answering systems focus on end-to-end performance, comparison of common components becomes difficult. To address this shortcoming, we present a quantitative evaluation of various passage retrieval algorithms for question answering, implemented in a framework called Pauchok. We present three important findings: Boolean querying schemes perform well in the question answering task. The performance differences between various passage retrieval algorithms vary with the choice of document retriever, which suggests significant interactions between document retrieval and passage retrieval. The best algorithms in our evaluation employ density-based measures for scoring query terms. Our results reveal future directions for passage retrieval and question answering.
language and technology conference | 2006
Gregory Marton; Alexey Radul
The TREC Definition and Relationship questions are evaluated on the basis of information nuggets that may be contained in system responses. Human evaluators provide informal descriptions of each nugget, and judgements (assignments of nuggets to responses) for each response submitted by participants. While human evaluation is the most accurate way to compare systems, approximate automatic evaluation becomes critical during system development.We present Nuggeteer, a new automatic evaluation tool for nugget-based tasks. Like the first such tool, Pourpre, Nuggeteer uses words in common between candidate answer and answer key to approximate human judgements. Unlike Pourpre, but like human assessors, Nuggeteer creates a judgement for each candidate-nugget pair, and can use existing judgements instead of guessing. This creates a more readily interpretable aggregate score, and allows developers to track individual nuggets through the variants of their system. Nuggeteer is quantitatively comparable in performance to Pourpre, and provides qualitatively better feedback to developers.
applied imagery pattern recognition workshop | 1999
Tapas Kanungo; Gregory Marton; Osama Bulbul
Numerous Optical Character Recognition (OCR) companies claim that their products have near-perfect recognition accuracy (close to 99.9%). In practice, however, these accuracy rates are rarely achieved. Most systems break down when the input document images are highly degraded, such as scanned images of carbon-copy documents, documents printed on low-quality paper, and documents that are n-th generation photocopies. Besides, the end user cannot compare the relative performances of the products because the various accuracy results are not reported on the same dataset.. In this article we report our evaluation results for two popular Arabic OCR products: (1) Sakhr OCR and (2) OmniPage for Arabic. In our evaluation we establish that the Sakhr OCR product has 15.47% lower page error rate relative to the OmniPage page error rate. The absolute page accuracy rates for Sakhr and Omnipage are 90.33% and 86.89% respectively. Our evaluation was performed using the SAIC Arabic image dataset, and we used only those pages for which both OCR systems produced output. A scatter-plot of the page accuracy-rate pairs reveals that Sakhr in general performs better on low-accuracy (degraded) pages. The scatter-plot visualization technique allows an algorithm developer to easily detect and analyze outliers in the results.
Lecture Notes in Computer Science | 2002
Boris Katz; Sue Felshin; Deniz Yuret; Ali Ibrahim; Jimmy J. Lin; Gregory Marton; Alton Jerome Mcfarland; Baris Temelkuran
text retrieval conference | 2003
Boris Katz; Jimmy J. Lin; Daniel Loreto; Wesley Hildebrandt; Matthew W. Bilotti; Sue Felshin; Aaron Fernandes; Gregory Marton; Federico Mora
applications of natural language to data bases | 2002
Boris Katz; Sue Felshin; Deniz Yuret; Ali Ibrahim; Jimmy J. Lin; Gregory Marton; Alton Jerome Mcfarland; Baris Temelkuran
text retrieval conference | 2006
Jimmy J. Lin; Aaron Fernandes; Boris Katz; Gregory Marton; Stefanie Tellex
text retrieval conference | 2004
Boris Katz; Matthew W. Bilotti; Sue Felshin; Aaron Fernandes; Wesley Hildebrandt; Roni Katzir; Jimmy J. Lin; Daniel Loreto; Gregory Marton; Federico Mora; Özlem Uzuner
text retrieval conference | 2006
Boris Katz; Gregory Marton; Sue Felshin; Daniel Loreto; Ben Lu; Federico Mora; Özlem Uzuner; Michael McGraw-Herdeg; Natalie Cheung; Alexey Radul; Yuan Kui Shen; Yuan Luo; Gabriel Zaccak
text retrieval conference | 2007
Boris Katz; Sue Felshin; Gregory Marton; Federico Mora; Yuan Kui Shen; Gabriel Zaccak; Ammar Ammar; Eric Eisner; Asli Turgut; L. Brown Westrick