Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Tom Kenter is active.

Publication


Featured researches published by Tom Kenter.


conference on information and knowledge management | 2015

Short Text Similarity with Word Embeddings

Tom Kenter; Maarten de Rijke

Determining semantic similarity between texts is important in many tasks in information retrieval such as search, query suggestion, automatic summarization and image finding. Many approaches have been suggested, based on lexical matching, handcrafted patterns, syntactic parse trees, external sources of structured semantic knowledge and distributional semantics. However, lexical features, like string matching, do not capture semantic similarity beyond a trivial level. Furthermore, handcrafted patterns and external sources of structured semantic knowledge cannot be assumed to be available in all circumstances and for all domains. Lastly, approaches depending on parse trees are restricted to syntactically well-formed texts, typically of one sentence in length. We investigate whether determining short text similarity is possible using only semantic features---where by semantic we mean, pertaining to a representation of meaning---rather than relying on similarity in lexical or syntactic representations. We use word embeddings, vector representations of terms, computed from unlabelled data, that represent terms in a semantic space in which proximity of vectors can be interpreted as semantic similarity. We propose to go from word-level to text-level semantics by combining insights from methods based on external sources of semantic knowledge with word embeddings. A novel feature of our approach is that an arbitrary number of word embedding sets can be incorporated. We derive multiple types of meta-features from the comparison of the word vectors for short text pairs, and from the vector means of their respective word embeddings. The features representing labelled short text pairs are used to train a supervised learning algorithm. We use the trained model at testing time to predict the semantic similarity of new, unlabelled pairs of short texts We show on a publicly available evaluation set commonly used for the task of semantic similarity that our method outperforms baseline methods that work under the same conditions.


meeting of the association for computational linguistics | 2016

Siamese CBOW: Optimizing Word Embeddings for Sentence Representations

Tom Kenter; Alexey Borisov; Maarten de Rijke

We present the Siamese Continuous Bag of Words (Siamese CBOW) model, a neural network for efficient estimation of high-quality sentence embeddings. Averaging the embeddings of words in a sentence has proven to be a surprisingly successful and efficient way of obtaining sentence embeddings. However, word embeddings trained with the methods currently available are not optimized for the task of sentence representation, and, thus, likely to be suboptimal. Siamese CBOW handles this problem by training word embeddings directly for the purpose of being averaged. The underlying neural network learns word embeddings by predicting, from a sentence representation, its surrounding sentences. We show the robustness of the Siamese CBOW model by evaluating it on 20 datasets stemming from a wide variety of sources.


conference on information and knowledge management | 2015

Ad Hoc Monitoring of Vocabulary Shifts over Time

Tom Kenter; Melvin Wevers; Pim Huijnen; Maarten de Rijke

Word meanings change over time. Detecting shifts in meaning for particular words has been the focus of much research recently. We address the complementary problem of monitoring shifts in vocabulary over time. That is, given a small seed set of words, we are interested in monitoring which terms are used over time to refer to the underlying concept denoted by the seed words. In this paper, we propose an algorithm for monitoring shifts in vocabulary over time, given a small set of seed terms. We use distributional semantic methods to infer a series of semantic spaces over time from a large body of time-stamped unstructured textual documents. We construct semantic networks of terms based on their representation in the semantic spaces and use graph-based measures to calculate saliency of terms. Based on the graph-based measures we produce ranked lists of terms that represent the concept underlying the initial seed terms over time as final output. As the task of monitoring shifting vocabularies over time for an ad hoc set of seed words is, to the best of our knowledge, a new one, we construct our own evaluation set. Our main contributions are the introduction of the task of ad hoc monitoring of vocabulary shifts over time, the description of an algorithm for tracking shifting vocabularies over time given a small set of seed words, and a systematic evaluation of results over a substantial period of time (over four decades). Additionally, we make our newly constructed evaluation set publicly available.


Information Processing and Management | 2015

Evaluating document filtering systems over time

Tom Kenter; Krisztian Balog; Maarten de Rijke

We propose a new way of measuring document filtering system performance over time.Performance is calculated per batch and a trend line is fitted to the results.Systems are compared by their performance at the end of the evaluation period.Important insights emerge by re-evaluating TREC KBA CCR runs of 2012 and 2013. Document filtering is a popular task in information retrieval. A stream of documents arriving over time is filtered for documents relevant to a set of topics. The distinguishing feature of document filtering is the temporal aspect introduced by the stream of documents. Document filtering systems, up to now, have been evaluated in terms of traditional metrics like (micro- or macro-averaged) precision, recall, MAP, nDCG, F1 and utility. We argue that these metrics do not capture all relevant aspects of the systems being evaluated. In particular, they lack support for the temporal dimension of the task. We propose a time-sensitive way of measuring performance of document filtering systems over time by employing trend estimation. In short, the performance is calculated for batches, a trend line is fitted to the results, and the estimated performance of systems at the end of the evaluation period is used to compare systems. We detail the application of our proposed trend estimation framework and examine the assumptions that need to hold for valid significance testing. Additionally, we analyze the requirements a document filtering metric has to meet and show that traditional macro-averaged true-positive-based metrics, like precision, recall and utility fail to capture essential information when applied in a batch setting. In particular, false positives returned in a batch for topics that are absent from the ground truth in that batch go unnoticed. This is a serious flaw as over-generation of a system might be overlooked this way. We propose a new metric, aptness, that does capture false positives. We incorporate this metric in an overall score and show that this new score does meet all requirements. To demonstrate the results of our proposed evaluation methodology, we analyze the runs submitted to the two most recent editions of a document filtering evaluation campaign. We re-evaluate the runs submitted to the Cumulative Citation Recommendation task of the 2012 and 2013 editions of the TREC Knowledge Base Acceleration track, and show that important new insights emerge.


european conference on information retrieval | 2017

Hierarchical Re-estimation of Topic Models for Measuring Topical Diversity

Hosein Azarbonyad; Mostafa Dehghani; Tom Kenter; Maarten Marx; Jaap Kamps; Maarten de Rijke

A high degree of topical diversity is often considered to be an important characteristic of interesting text documents. A recent proposal for measuring topical diversity identifies three elements for assessing diversity: words, topics, and documents as collections of words. Topic models play a central role in this approach. Using standard topic models for measuring diversity of documents is suboptimal due to generality and impurity. General topics only include common information from a background corpus and are assigned to most of the documents in the collection. Impure topics contain words that are not related to the topic; impurity lowers the interpretability of topic models and impure topics are likely to get assigned to documents erroneously. We propose a hierarchical re-estimation approach for topic models to combat generality and impurity; the proposed approach operates at three levels: words, topics, and documents. Our re-estimation approach for measuring documents’ topical diversity outperforms the state of the art on PubMed dataset which is commonly used for diversity experiments.


Lecture Notes in Computer Science | 2014

Advances in information retrieval: 36th European Conference on IR Research, ECIR 2014, Amsterdam, The Netherlands, April 13-16, 2014: proceedings

de M. Rijke; Tom Kenter; de Arjen Vries; ChengXiang Zhai; de Franciska Jong; Kira Radinsky; Katja Hofmann

This is a summary of the keynote talk delivered by the winner of the Karen Spärck-Jones Award at ECIR 2014. A long standing challenge in Web search is how to accurately determine the intention behind a searchers query, which is needed to rank, organize, and present results most effectively. The difficulty is that users often do not (or cannot) provide sufficient information about their goals. As this talk with show, it is nevertheless possible to read their intentions through clues revealed by behavior, such as the amount of attention paid to a document or a text fragment. I will overview the approaches that have emerged for acquiring and mining behavioral data for inferring search intent, ranging from robust models of click data in the aggregate, to modeling fine-grained user interactions such as mouse cursor movements in the searchers browser. The latter can also be used to measure the searchers attention “in the wild, with granularity approaching that of using eye tracking equipment in the laboratory. The resulting techniques and models have already shown noteworthy improvements for search tasks such as ranking, relevance estimation, and result summary generation, and have applications to other domains, such as psychology, neurology, and online education. Real-Time Search at Twitter


Theory and Applications of Categories | 2012

Context-Based Entity Linking - University of Amsterdam at TAC 2012

David Graus; Tom Kenter; Marc Bron; Edgar Meij; Maarten de Rijke


arXiv: Computation and Language | 2017

Attentive Memory Networks: Efficient Machine Reading for Conversational Search.

Tom Kenter; Maarten de Rijke


international acm sigir conference on research and development in information retrieval | 2017

Neural Networks for Information Retrieval

Tom Kenter; Alexey Borisov; Christophe Van Gysel; Mostafa Dehghani; Maarten de Rijke; Bhaskar Mitra


text retrieval conference | 2013

Filtering Documents over Time on Evolving Topics - The University of Amsterdam at TREC 2013 KBA CCR.

Tom Kenter

Collaboration


Dive into the Tom Kenter's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

M. de Rijke

University of Amsterdam

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kira Radinsky

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Arjen P. de Vries

Radboud University Nijmegen

View shared research outputs
Top Co-Authors

Avatar

David Graus

University of Amsterdam

View shared research outputs
Top Co-Authors

Avatar

Franciska de Jong

Erasmus University Rotterdam

View shared research outputs
Researchain Logo
Decentralizing Knowledge