Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jaime G. Carbonell is active.

Publication


Featured researches published by Jaime G. Carbonell.


international acm sigir conference on research and development in information retrieval | 1998

The Use of MMR and Diversity-Based Reranking for Reodering Documents and Producing Summaries

Jaime G. Carbonell; Jade Goldstein

This paper presents a method for combining query-relevance with information-novelty in the context of text retrieval and summarization. The Maximal Marginal Relevance (MMR) criterion strives to reduce redundancy while maintaining query relevance in re-ranking retrieved documents and in selecting appropriate passages for text summarization. Preliminary results indicate some benefits for MMR diversity ranking in document retrieval and in single document summarization. The latter are borne out by the recent results of the SUMMAC conference in the evaluation of summarization systems. However, the clearest advantage is demonstrated in constructing non-redundant multi-document summaries, where MMR results are clearly superior to non-MMR passage selection.


international acm sigir conference on research and development in information retrieval | 1998

A study of retrospective and on-line event detection

Yiming Yang; Thomas Pierce; Jaime G. Carbonell

This paper investigates the use and extension of text retrieval and clustering techniques for event detection. The task is to automatically detect novel events from a temporally-ordered stream of news stories, either retrospectively or as the stories arrive. We applied hierarchical and non-hierarchical document clustering algorithms to a corpus of 15,836 stories, focusing on the exploitation of both content and temporal information. We found the resulting cluster hierarchies highly informative for retrospective detection of previously unidentified events, effectively supporting both query-free and query-driven retrieval. We also found that temporal distribution patterns of document clusters provide useful information for improvement in both retrospective detection and on-line detection of novel events. In an evaluation using manually labelled events to judge the system-detected events, we obtained a result of 82% in the Fl measure for retrospective detection, and a Fl value of 42% for on-line detection.


international acm sigir conference on research and development in information retrieval | 1999

Summarizing text documents: sentence selection and evaluation metrics

Jade Goldstein; Mark Kantrowitz; Vibhu O. Mittal; Jaime G. Carbonell

Human-quality text summarization systems are di cult to design, and even more di cult to evaluate, in part because documents can di er along several dimensions, such as length, writing style and lexical usage. Nevertheless, certain cues can often help suggest the selection of sentences for inclusion in a summary. This paper presents our analysis of news-article summaries generated by sentence selection. Sentences are ranked for potential inclusion in the summary using a weighted combination of statistical and linguistic features. The statistical features were adapted from standard IR methods. The potential linguistic ones were derived from an analysis of news-wire summaries. To evaluate these features we use a normalized version of precision-recall curves, with a baseline of random sentence selection, as well as analyze the properties of such a baseline. We illustrate our discussions with empirical results showing the importance of corpus-dependent baseline summarization standards, compression ratios and carefully crafted long queries.


Machine Learning#R##N#An Artificial Intelligence Approach, Volume I | 1983

Learning by analogy : formulating and generalizing plans from past experience

Jaime G. Carbonell

Analogical reasoning is a powerful mechanism for exploiting past experience in planning and problem solving. This chapter outlines a theory of analogical problem solving based on an extension to means-ends analysis. An analogical transformation process is developed to extract knowledge from past successful problem-solving situations that bear a strong similarity to the current problem. Then, the investigation focuses on exploiting and extending the analogical reasoning model to generate useful exemplary solutions to related problems from which more general plans can be induced and refined. Starting with a general analogical inference engine, problem-solving experience is, in essence, compiled incrementally into effective procedures that solve various classes of problems in an increasingly reliable and direct manner.


Journal of Experimental and Theoretical Artificial Intelligence | 1995

Integrating planning and learning: the PRODIGY architecture

Manuela M. Veloso; Jaime G. Carbonell; M. Alicia Pérez; Daniel Borrajo; Eugene Fink; Jim Blythe

Abstract Planning is a complex reasoning task that is well suited for the study of improving performance and knowledge by learning, i.e. by accumulation and interpretation of planning experience. PRODIGY is an architecture that integrates planning with multiple learning mechanisms. Learning occurs at the planners decision points and integration in PRODIGY is achieved via mutually interpretable knowledge structures. This article describes the PRODIGY planner, briefly reports on several learning modules developed earlier along the project, and presents in more detail two recently explored methods to learn to generate plans of better quality. We introduce the techniques, illustrate them with comprehensive examples, and show preliminary empirical results. The article also includes a retrospective discussion of the characteristics of the overall PRODIGY architecture and discusses their evolution within the goal of the project of building a large and robust integrated planning and learning system.


siam international conference on data mining | 2010

Temporal Collaborative Filtering with Bayesian Probabilistic Tensor Factorization.

Liang Xiong; Xi Chen; Tzu-Kuo Huang; Jeff G. Schneider; Jaime G. Carbonell

Real-world relational data are seldom stationary, yet traditional collaborative filtering algorithms generally rely on this assumption. Motivated by our sales prediction problem, we propose a factor-based algorithm that is able to take time into account. By introducing additional factors for time, we formalize this problem as a tensor factorization with a special constraint on the time dimension. Further, we provide a fully Bayesian treatment to avoid tuning parameters and achieve automatic model complexity control. To learn the model we develop an efficient sampling procedure that is capable of analyzing large-scale data sets. This new algorithm, called Bayesian Probabilistic Tensor Factorization (BPTF), is evaluated on several real-world problems including sales prediction and movie recommendation. Empirical results demonstrate the superiority of our temporal model.


north american chapter of the association for computational linguistics | 2000

Multi-document summarization by sentence extraction

Jade Goldstein; Vibhu O. Mittal; Jaime G. Carbonell; Mark Kantrowitz

This paper discusses a text extraction approach to multi-document summarization that builds on single-document summarization methods by using additional, available information about the document set as a whole and the relationships between the documents. Multi-document summarization differs from single in that the issues of compression, speed, redundancy and passage selection are critical in the formation of useful summaries. Our approach addresses these issues by using domain-independent techniques based mainly on fast, statistical processing, a metric for reducing redundancy and maximizing diversity in the selected passages, and a modular framework to allow easy parameterization for different genres, corpora characteristics and user requirements.


Transactions on Rough Sets | 1983

AN OVERVIEW OF MACHINE LEARNING

Jaime G. Carbonell; Ryszard S. Michalski; Tom M. Mitchell

Learning is a many-faceted phenomenon. Learning processes include the acquisition of new declarative knowledge, the development of motor and cognitive skills through instruction or practice, the organization of new knowledge into general, effective representations, and the discovery of new facts and theories through observation and experimentation. Since the inception of the computer era, researchers have been striving to implant such capabilities in computers. Solving this problem has been, and remains, a most challenging and fascinating long-range goal in artificial intelligence (AI). The study and computer modeling of learning processes in their multiple manifestations constitutes the subject matter of machine learning.


Artificial Intelligence | 1989

Explanation-based learning: a problem solving perspective

Steven Minton; Jaime G. Carbonell; Craig A. Knoblock; Daniel R. Kuokka; Oren Etzioni; Yolanda Gil

Abstract This article outlines explanation-based learning (EBL) and its role in improving problem solving performance through experience. Unlike inductive systems, which learn by abstracting common properties from multiple examples, EBL systems explain why a particular example is an instance of a concept. The explanations are then converted into operational recognition rules. In essence, the EBL approach is analytical and knowledge-intensive, whereas inductive methods are empirical and knowledge-poor. This article focuses on extensions of the basic EBL method and their integration with the prodigy problem solving system. prodigy s EBL method is specifically designed to acquire search control rules that are effective in reducing total search time for complex task domains. Domain-specific search control rules are learned from successful problem solving decisions, costly failures, and unforeseen goal interactions. The ability to specify multiple learning strategies in a declarative manner enables EBL to serve as a general technique for performance improvement. prodigy s EBL method is analyzed, illustrated with several examples and performance results, and compared with other methods for integrating EBL and problem solving.


knowledge discovery and data mining | 2002

Topic-conditioned novelty detection

Yiming Yang; Jian Zhang; Jaime G. Carbonell; Chun Jin

Automated detection of the first document reporting each new event in temporally-sequenced streams of documents is an open challenge. In this paper we propose a new approach which addresses this problem in two stages: 1) using a supervised learning algorithm to classify the on-line document stream into pre-defined broad topic categories, and 2) performing topic-conditioned novelty detection for documents in each topic. We also focus on exploiting named-entities for event-level novelty detection and using feature-based heuristics derived from the topic histories. Evaluating these methods using a set of broadcast news stories, our results show substantial performance gains over the traditional one-level approach to the novelty detection problem.

Collaboration


Dive into the Jaime G. Carbonell's collaboration.

Top Co-Authors

Avatar

Yiming Yang

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Eugene Fink

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Anatole Gershman

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Ralf D. Brown

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Lori S. Levin

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Alon Lavie

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Peter J. Jansen

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Yan Liu

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Philip J. Hayes

Carnegie Mellon University

View shared research outputs
Researchain Logo
Decentralizing Knowledge