Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Peter Wilkins is active.

Publication


Featured researches published by Peter Wilkins.


international acm sigir conference on research and development in information retrieval | 2010

Properties of optimally weighted data fusion in CBMIR

Peter Wilkins; Alan F. Smeaton; Paul Ferguson

Content-Based Multimedia Information Retrieval (CBMIR) systems which leverage multiple retrieval experts (En) often employ a weighting scheme when combining expert results through data fusion. Typically however a query will comprise multiple query images (Im) leading to potentially N × M weights to be assigned. Because of the large number of potential weights, existing approaches impose a hierarchy for data fusion, such as uniformly combining query image results from a single retrieval expert into a single list and then weighting the results of each expert. In this paper we will demonstrate that this approach is sub-optimal and leads to the poor state of CBMIR performance in benchmarking evaluations. We utilize an optimization method known as Coordinate Ascent to discover the optimal set of weights (|En| ⋅ |Im|) which demonstrates a dramatic difference between known results and the theoretical maximum. We find that imposing common combinatorial hierarchies for data fusion will half the optimal performance that can be achieved. By examining the optimal weight sets at the topic level, we observe that approximately 15% of the weights (from set |En| ⋅ |Im|) for any given query, are assigned 70%-82% of the total weight mass for that topic. Furthermore we discover that the ideal distribution of weights follows a log-normal distribution. We find that we can achieve up to 88% of the performance of fully optimized query using just these 15% of the weights. Our investigation was conducted on TRECVID evaluations 2003 to 2007 inclusive and ImageCLEFPhoto 2007, totalling 181 search topics optimized over a combined collection size of 661,213 images and 1,594 topic images.


european conference on information retrieval | 2005

Físréal: a low cost terabyte search engine

Paul Ferguson; Cathal Gurrin; Peter Wilkins; Alan F. Smeaton

In this poster we describe the development of a distributed search engine, referred to as Fisreal, which utilises inexpensive workstations, yet attains fast retrieval performance for Terabyte-sized collections. We also discuss the process of leveraging additional meaning from the structure of HTML, as well as the use of anchor text documents to increase retrieval performance.


Signal Processing-image Communication | 2007

Inexpensive fusion methods for enhancing feature detection

Peter Wilkins; Tomasz Adamek; Noel E. O'Connor; Alan F. Smeaton

Recent successful approaches to high-level feature detection in image and video data have treated the problem as a pattern classification task. These typically leverage the techniques learned from statistical machine learning, coupled with ensemble architectures that create multiple feature detection models. Once created, co-occurrence between learned features can be captured to further boost performance. At multiple stages throughout these frameworks, various pieces of evidence can be fused together in order to boost performance. These approaches whilst very successful are computationally expensive, and depending on the task, require the use of significant computational resources. In this paper we propose two fusion methods that aim to combine the output of an initial basic statistical machine learning approach with a lower-quality information source, in order to gain diversity in the classified results whilst requiring only modest computing resources. Our approaches, validated experimentally on TRECVid data, are designed to be complementary to existing frameworks and can be regarded as possible replacements for the more computationally expensive combination strategies used elsewhere.


european conference on information retrieval | 2006

Automatic determination of feature weights for multi-feature CBIR

Peter Wilkins; Paul Ferguson; Cathal Gurrin; Alan F. Smeaton

Image and video retrieval are both currently dominated by approaches which combine the outputs of several different representations or features. The ways in which the combination can be done is an established research problem in content-based image retrieval (CBIR). These approaches vary from image clustering through to semantic frameworks and mid-level visual features to ultimately determine sets of relative weights for the non-linear combination of features. Simple approaches to determining these weights revolve around executing a standard set of queries with known relevance judgements on some form of training data and is iterative in nature. Whilst successful, this requires both training data and human intervention to derive the optimal weights.


cross language evaluation forum | 2009

Document expansion for text-based image retrieval at CLEF 2009

Jinming Min; Peter Wilkins; Johannes Leveling; Gareth J. F. Jones

In this paper, we describe and analyze our participation in the WikipediaMM task at CLEF 2009. Our main efforts concern the expansion of the image metadata from the Wikipedia abstracts collection - DBpedia. In our experiments, we use the Okapi feedback algorithm for document expansion. Compared with our text retrieval baseline, our best document expansion RUN improves MAP by 17.89%. As one of our conclusions, document expansion from external resource can play an effective factor in the image metadata retrieval task.


content based multimedia indexing | 2007

Inexpensive Fusion Methods for Enhancing Feature Detection

Peter Wilkins; Tomasz Adamek; Noel E. O'Connor; Alan F. Smeaton

In this paper we present two fusion methods for the task of high-level feature detection in multimedia content. Successful approaches to high-level feature detection typically leverage the techniques learned from Machine Learning utilized through ensemble architectures to achieve strong performance. However these approaches whilst successful are computationally expensive, and depending on the task require the use of significant computational resources. We propose two fusion methods that aim to combine the output of an initial basic machine learning approach with a lower-quality information source in order to gain diversity in the classified results whilst only requiring modest computing resources.


international acm sigir conference on research and development in information retrieval | 2005

Top subset retrieval on large collections using sorted indices

Paul Ferguson; Alan F. Smeaton; Cathal Gurrin; Peter Wilkins

In this poster we describe alternative inverted index structures that reduce the time required to process queries, produce a higher query throughput and still return high quality results to the end user. We give results based upon the TREC Terabyte dataset showing improvements that these indices give in terms of effectiveness and efficiency.


conference on image and video retrieval | 2009

DCU collaborative video search system

Colum Foley; Peter Wilkins; Alan F. Smeaton

This paper describes the Dublin City University collaborative video search system. The system is a development on our previous years VideOlympics submission which will explore the notion of division of labour and sharing of knowledge across collaborating users engaged in a shared search. Division of labour and sharing of knowledge across collaborating searchers is realised through system-mediated coordination of the search.


cross language evaluation forum | 2008

Diversity in image retrieval: DCU at ImageCLEFPhoto 2008

Neil O'Hare; Peter Wilkins; Cathal Gurrin; Eamonn Newman; Gareth J. F. Jones; Alan F. Smeaton

DCU participated in the ImageCLEF 2008 photo retrieval task, which aimed to evaluate diversity in Image Retrieval, submitting runs for both the English and Random language annotation conditions. Our approaches used text-based and image-based retrieval to give baseline runs, with the the highest-ranked images from these baseline runs clustered using K-Means clustering of the text annotations, with representative images from each cluster ranked for the final submission. For random language annotations, we compared results from translated runs with untranslated runs. Our results show that combining image and text outperforms text alone and image alone, both for general retrieval performance and for diversity. Our baseline image and text runs give our best overall balance between retrieval and diversity; indeed, our baseline text and image run was the 2nd best automatic run for ImageCLEF 2008 Photographic Retrieval task. We found that clustering consistently gives a large improvement in diversity performance over the baseline, unclustered results, while degrading retrieval performance. Pseudo relevance feedback consistently improved retrieval, but always at the cost of diversity. We also found that the diversity of untranslated random runs was quite close to that of translated random runs, indicating that for this dataset at least, if diversity is our main concern it may not be necessary to translate the image annotations.


multimedia information retrieval | 2006

Using score distributions for query-time fusion in multimediaretrieval

Peter Wilkins; Paul Ferguson; Alan F. Smeaton

Collaboration


Dive into the Peter Wilkins's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jinming Min

Dublin City University

View shared research outputs
Researchain Logo
Decentralizing Knowledge