Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Gaurav Baruah is active.

Publication


Featured researches published by Gaurav Baruah.


international acm sigir conference on research and development in information retrieval | 2014

The effect of expanding relevance judgements with duplicates

Gaurav Baruah; Adam Roegiest; Mark D. Smucker

We examine the effects of expanding a judged set of sentences with their duplicates from a corpus. Including new sentences that are exact duplicates of the previously judged sentences may allow for better estimation of performance metrics and enhance the reusability of a test collection. We perform experiments in context of the Temporal Summarization Track at TREC 2013. We find that adding duplicate sentences to the judged set does not significantly affect relative system performance. However, we do find statistically significant changes in the performance of nearly half the systems that participated in the Track. We recommend adding exact duplicate sentences to the set of relevance judgements in order to obtain a more accurate estimate of system performance.


international acm sigir conference on research and development in information retrieval | 2015

Evaluating Streams of Evolving News Events

Gaurav Baruah; Mark D. Smucker; Charles L. A. Clarke

People track news events according to their interests and available time. For a major event of great personal interest, they might check for updates several times an hour, taking time to keep abreast of all aspects of the evolving event. For minor events of more marginal interest, they might check back once or twice a day for a few minutes to learn about the most significant developments. Systems generating streams of updates about evolving events can improve user performance by appropriately filtering these updates, making it easy for users to track events in a timely manner without undue information overload. Unfortunately, predicting user performance on these systems poses a significant challenge. Standard evaluation methodology, designed for Web search and other adhoc retrieval tasks, adapts poorly to this context. In this paper, we develop a simple model that simulates users checking the system from time to time to read updates. For each simulated user, we generate a trace of their activities alternating between away times and reading times. These traces are then applied to measure system effectiveness. We test our model using data from the TREC 2013 Temporal Summarization Track (TST) comparing it to the effectiveness measures used in that track. The primary TST measure corresponds most closely with a modeled user that checks back once a day on average for an average of one minute. Users checking more frequently for longer times may view the relative performance of participating systems quite differently. In light of this sensitivity to user behavior, we recommend that future experiments be built around clearly stated assumptions regarding user interfaces and access patterns, with effectiveness measures reflecting these assumptions.


conference on information and knowledge management | 2016

Optimizing Nugget Annotations with Active Learning

Gaurav Baruah; Haotian Zhang; Rakesh Guttikonda; Jimmy J. Lin; Mark D. Smucker; Olga Vechtomova

Nugget-based evaluations, such as those deployed in the TREC Temporal Summarization and Question Answering tracks, require human assessors to determine whether a nugget is present in a given piece of text. This process, known as nugget annotation, is labor-intensive. In this paper, we present two active learning techniques that prioritize the sequence in which candidate nugget/sentence pairs are presented to an assessor, based on the likelihood that the sentence contains a nugget. Our approach builds on the recognition that nugget annotation is similar to high-recall retrieval, and we adapt proven existing solutions. Simulation experiments with four existing TREC test collections show that our techniques yield far more matches for a given level of effort than baselines that are typically deployed in previous nugget-based evaluations.


IEEE Internet Computing | 2016

Searching from Mars

Jimmy J. Lin; Charles L. A. Clarke; Gaurav Baruah

How would you search from Mars? Its the user model, stupid!


international conference on the theory of information retrieval | 2015

Pooling for User-Oriented Evaluation Measures

Gaurav Baruah; Adam Roegiest; Mark D. Smucker

Traditional TREC-style pooling methodology relies on using predicted relevance by systems to select documents for judgment. This coincides with typical search behaviour (e.g., web search). In the case of temporally ordered streams of documents, the order that users encounter documents is in this temporal order and not some predetermined rank order. We investigate a user oriented pooling methodology focusing on the documents that simulated users would likely read in such temporally ordered streams. Under this user model, many of the relevant documents found in the TREC 2013 Temporal Summarization Tracks pooling effort would never be read. Not only does our pooling strategy focus on pooling documents that will be read by (simulated) users, the resultant pools are different from the standard TREC pools.


conference on information and knowledge management | 2017

A Comparison of Nuggets and Clusters for Evaluating Timeline Summaries

Gaurav Baruah; Richard McCreadie; Jimmy J. Lin


arXiv: Information Retrieval | 2017

Exploring the Effectiveness of Convolutional Neural Networks for Answer Selection in End-to-End Question Answering.

Royal Sequiera; Gaurav Baruah; Zhucheng Tu; Salman Mohammed; Jinfeng Rao; Haotian Zhang; Jimmy J. Lin


international conference on the theory of information retrieval | 2017

The Pareto Frontier of Utility Models as a Framework for Evaluating Push Notification Systems

Gaurav Baruah; Jimmy J. Lin


international acm sigir conference on research and development in information retrieval | 2017

On the Reusability of "Living Labs" Test Collections: A Case Study of Real-Time Summarization

Luchen Tan; Gaurav Baruah; Jimmy J. Lin


international acm sigir conference on research and development in information retrieval | 2018

Klick Labs at CL-SciSumm 2018.

Gaurav Baruah; Maheedhar Kolla

Collaboration


Dive into the Gaurav Baruah's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Luchen Tan

University of Waterloo

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge