Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Evangelos Kanoulas is active.

Publication


Featured researches published by Evangelos Kanoulas.


international conference on data engineering | 2006

Finding Fastest Paths on A Road Network with Speed Patterns

Evangelos Kanoulas; Yang Du; Tian Xia; Donghui Zhang

This paper proposes and solves the Time-Interval All Fastest Path (allFP) query. Given a user-defined leaving or arrival time interval I, a source node s and an end node e, allFP asks for a set of all fastest paths from s to e, one for each sub-interval of I. Note that the query algorithm should find a partitioning of I into sub-intervals. Existing methods can only be used to solve a very special case of the problem, when the leaving time is a single time instant. A straightforward solution to the allFP query is to run existing methods many times, once for every time instant in I. This paper proposes a solution based on novel extensions to the A* algorithm. Instead of expanding the network many times, we expand once. The travel time on a path is kept as a function of leaving time. Methods to combine travel-time functions are provided to expand a path. A novel lower-bound estimator for travel time is proposed. Performance results reveal that our method is more efficient and more accurate than the discrete-time approach.


international acm sigir conference on research and development in information retrieval | 2008

Evaluation over thousands of queries

Ben Carterette; Virgiliu Pavlu; Evangelos Kanoulas; Javed A. Aslam; James Allan

Information retrieval evaluation has typically been performed over several dozen queries, each judged to near-completeness. There has been a great deal of recent work on evaluation over much smaller judgment sets: how to select the best set of documents to judge and how to estimate evaluation measures when few judgments are available. In light of this, it should be possible to evaluate over many more queries without much more total judging effort. The Million Query Track at TREC 2007 used two document selection algorithms to acquire relevance judgments for more than 1,800 queries. We present results of the track, along with deeper analysis: investigating tradeoffs between the number of queries and number of judgments shows that, up to a point, evaluation over more queries with fewer judgments is more cost-effective and as reliable as fewer queries with more judgments. Total assessor effort can be reduced by 95% with no appreciable increase in evaluation errors.


international acm sigir conference on research and development in information retrieval | 2010

Extending average precision to graded relevance judgments

Stephen E. Robertson; Evangelos Kanoulas; Emine Yilmaz

Evaluation metrics play a critical role both in the context of comparative evaluation of the performance of retrieval systems and in the context of learning-to-rank (LTR) as objective functions to be optimized. Many different evaluation metrics have been proposed in the IR literature, with average precision (AP) being the dominant one due a number of desirable properties it possesses. However, most of these measures, including average precision, do not incorporate graded relevance. In this work, we propose a new measure of retrieval effectiveness, the Graded Average Precision (GAP). GAP generalizes average precision to the case of multi-graded relevance and inherits all the desirable characteristics of AP: it has a nice probabilistic interpretation, it approximates the area under a graded precision-recall curve and it can be justified in terms of a simple but moderately plausible user model. We then evaluate GAP in terms of its informativeness and discriminative power. Finally, we show that GAP can reliably be used as an objective metric in learning to rank by illustrating that optimizing for GAP using SoftRank and LambdaRank leads to better performing ranking functions than the ones constructed by algorithms tuned to optimize for AP or NDCG even when using AP or NDCG as the test metrics.


international acm sigir conference on research and development in information retrieval | 2011

Evaluating multi-query sessions

Evangelos Kanoulas; Ben Carterette; Paul D. Clough; Mark Sanderson

The standard system-based evaluation paradigm has focused on assessing the performance of retrieval systems in serving the best results for a single query. Real users, however, often begin an interaction with a search engine with a sufficiently under-specified query that they will need to reformulate before they find what they are looking for. In this work we consider the problem of evaluating retrieval systems over test collections of multi-query sessions. We propose two families of measures: a model-free family that makes no assumption about the users behavior over a session, and a model-based family with a simple model of user interactions over the session. In both cases we generalize traditional evaluation metrics such as average precision to multi-query session evaluation. We demonstrate the behavior of the proposed metrics by using the new TREC 2010 Session track collection and simulations over the TREC-9 Query track collection.


conference on information and knowledge management | 2011

Simulating simple user behavior for system effectiveness evaluation

Ben Carterette; Evangelos Kanoulas; Emine Yilmaz

Information retrieval effectiveness evaluation typically takes one of two forms: batch experiments based on static test collections, or lab studies measuring actual users interacting with a system. Test collection experiments are sometimes viewed as introducing too many simplifying assumptions to accurately predict the usefulness of a system to its users. As a result, there is great interest in creating test collections and measures that better model user behavior. One line of research involves developing measures that include a parameterized user model; choosing a parameter value simulates a particular type of user. We propose that these measures offer an opportunity to more accurately simulate the variance due to user behavior, and thus to analyze system effectiveness to a simulated user population. We introduce a Bayesian procedure for producing sampling distributions from click data, and show how to use statistical tools to quantify the effects of variance due to parameter selection.


international acm sigir conference on research and development in information retrieval | 2009

Document selection methodologies for efficient and effective learning-to-rank

Javed A. Aslam; Evangelos Kanoulas; Virgiliu Pavlu; Stefan Savev; Emine Yilmaz

Learning-to-rank has attracted great attention in the IR community. Much thought and research has been placed on query-document feature extraction and development of sophisticated learning-to-rank algorithms. However, relatively little research has been conducted on selecting documents for learning-to-rank data sets nor on the effect of these choices on the efficiency and effectiveness of learning-to-rank algorithms. In this paper, we employ a number of document selection methodologies, widely used in the context of evaluation--depth-k pooling, sampling (infAP, statAP), active-learning (MTC), and on-line heuristics (hedge). Certain methodologies, e.g. sampling and active-learning, have been shown to lead to efficient and effective evaluation. We investigate whether they can also enable efficient and effective learning-to-rank. We compare them with the document selection methodology used to create the LETOR datasets. Further, all of the utilized methodologies are different in nature, and thus they construct training data sets with different properties, such as the proportion of relevant documents in the data or the similarity among them. We study how such properties affect the efficiency, effectiveness, and robustness of learning-to-rank collections.


Physics in Medicine and Biology | 2007

Derivation of the tumor position from external respiratory surrogates with periodical updating of the internal/external correlation

Evangelos Kanoulas; Javed A. Aslam; G Sharp; R Berbeco; Seiko Nishioka; Hiroki Shirato; S Jiang

In this work we develop techniques that can derive the tumor position from external respiratory surrogates (abdominal surface motion) through periodically updated internal/external correlation. A simple linear function is used to express the correlation between the tumor and surrogate motion. The function parameters are established during a patient setup session with the tumor and surrogate positions simultaneously measured at a 30 Hz rate. During treatment, the surrogate position, constantly acquired at 30 Hz, is used to derive the tumor position. Occasionally, a pair of radiographic images is acquired to enable the updating of the linear correlation function. Four update methods, two aggressive and two conservative, are investigated: (A1) shift line through the update point; (A2) re-fit line through the update point; (C1) re-fit line with extra weight to the update point; (C2) minimize the distances to the update point and previous line fit point. In the present study of eight lung cancer patients, tumor and external surrogate motion demonstrate a high degree of linear correlation which changes dynamically over time. It was found that occasionally updating the correlation function leads to more accurate predictions than using external surrogates alone. In the case of high imaging rates during treatment (greater than 2 Hz) the aggressive update methods (A1 and A2) are more accurate than the conservative ones (C1 and C2). The opposite is observed in the case of low imaging rates.


conference on information and knowledge management | 2009

Empirical justification of the gain and discount function for nDCG

Evangelos Kanoulas; Javed A. Aslam

The nDCG measure has proven to be a popular measure of retrieval effectiveness utilizing graded relevance judgments. However, a number of different instantiations of nDCG exist, depending on the arbitrary definition of the gain and discount functions used (1) to dictate the relative value of documents of different relevance grades and (2) to weight the importance of gain values at different ranks, respectively. In this work we discuss how to empirically derive a gain and discount function that optimizes the efficiency or stability of nDCG. First, we describe a variance decomposition analysis framework and an optimization procedure utilized to find the efficiency- or stability-optimal gain and discount functions. Then we use TREC data sets to compare the optimal gain and discount functions to the ones that have appeared in the IR literature with respect to (a) the efficiency of the evaluation, (b) the induced ranking of systems, and (c) the discriminative power of the resulting nDCG measure.


international acm sigir conference on research and development in information retrieval | 2012

On per-topic variance in IR evaluation

Stephen E. Robertson; Evangelos Kanoulas

We explore the notion, put forward by Cormack & Lynam and Robertson, that we should consider a document collection used for Cranfield-style experiments as a sample from some larger population of documents. In this view, any per-topic metric (such as average precision) should be regarded as an estimate of that metrics true value for that topic in the full population, and therefore as carrying its own per-topic variance or estimate precision or noise. As in the two mentioned papers, we explore this notion by simulating other samples from the same large population. We investigate different ways of performing this simulation. One use of this analysis is to refine the notion of statistical significance of a difference between two systems (in most such analyses, each per-topic measurement is treated as equally precise). We propose a mixed-effects model method to measure significance, and compare it experimentally with the traditional t-test.


extending database technology | 2004

A Framework for Access Methods for Versioned Data

Betty Salzberg; Linan Jiang; David B. Lomet; Manuel Barrena; Jing Shan; Evangelos Kanoulas

This paper presents a framework for understanding and constructing access methods for versioned data. Records are associated with version ranges in a version tree. A minimal representation for the end set of a version range is given. We show how, within a page, a compact representation of a record can be made using start version of the version range only. Current-version splits, version-and-key splits and consolidations are explained. These operations preserve an invariant which allows visiting only one page at each level of the access method when doing exact-match search (no backtracking). Splits and consolidations also enable efficient stabbing queries by clustering data alive at a given version into a small number of data pages. Last, we survey the methods in the literature to show in what ways they conform or do not conform to our framework. These methods include temporal access methods, branched versioning access methods and spatio-temporal access methods. Our contribution is not to create a new access method but to bring to light fundamental properties of version-splitting access methods and to provide a blueprint for future versioned access methods. In addition, we have not made the unrealistic assumption that transactions creating a new version make only one update, and have shown how to treat multiple updates.

Collaboration


Dive into the Evangelos Kanoulas's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Emine Yilmaz

University College London

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dan Li

University of Amsterdam

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge