Ken Chatfield
University of Oxford
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Ken Chatfield.
british machine vision conference | 2014
Ken Chatfield; Karen Simonyan; Andrea Vedaldi; Andrew Zisserman
The latest generation of Convolutional Neural Networks (CNN) have achieved impressive results in challenging benchmarks on image recognition and object detection, significantly raising the interest of the community in these methods. Nevertheless, it is still unclear how different CNN methods compare with each other and with previous state-of-the-art shallow representations such as the Bag-of-Visual-Words and the Improved Fisher Vector. This paper conducts a rigorous evaluation of these new techniques, exploring different deep architectures and comparing them on a common ground, identifying and disclosing important implementation details. We identify several useful properties of CNN-based representations, including the fact that the dimensionality of the CNN output layer can be reduced significantly without having an adverse effect on performance. We also identify aspects of deep and shallow methods that can be successfully shared. In particular, we show that the data augmentation techniques commonly applied to CNN-based methods can also be applied to shallow methods, and result in an analogous performance boost. Source code and models to reproduce the experiments in the paper is made publicly available.
british machine vision conference | 2011
Ken Chatfield; Victor S. Lempitsky; Andrea Vedaldi; Andrew Zisserman
A large number of novel encodings for bag of visual words models have been proposed in the past two years to improve on the standard histogram of quantized local features. Examples include locality-constrained linear encoding [23], improved Fisher encoding [17], super vector encoding [27], and kernel codebook encoding [20]. While several authors have reported very good results on the challenging PASCAL VOC classification data by means of these new techniques, differences in the feature computation and learning algorithms, missing details in the description of the methods, and different tuning of the various components, make it impossible to compare directly these methods and hard to reproduce the results reported. This paper addresses these shortcomings by carrying out a rigorous evaluation of these new techniques by: (1) fixing the other elements of the pipeline (features, learning, tuning); (2) disclosing all the implementation details, and (3) identifying both those aspects of each method which are particularly important to achieve good performance, and those aspects which are less critical. This allows a consistent comparative analysis of these encoding methods. Several conclusions drawn from our analysis cannot be inferred from the original publications.
international conference on computer vision | 2009
Ken Chatfield; James Philbin; Andrew Zisserman
We present an efficient object retrieval system based on the identification of abstract deformable ‘shape’ classes using the self-similarity descriptor of Shechtman and Irani [13]. Given a user-specified query object, we retrieve other images which share a common ‘shape’ even if their appearance differs greatly in terms of colour, texture, edges and other common photometric properties. In order to use the self-similarity descriptor for efficient retrieval we make three contributions: (i) we sparsify the descriptor points by locating discriminative regions within each image, thus reducing the computational expense of shape matching; (ii) we extend [13] to enable matching despite changes in scale; and (iii) we show that vector quantizing the descriptor does not inhibit performance, thus providing the basis of a large-scale shape-based retrieval system using a bag-of-visual-words approach. Performance is demonstrated on the challenging ETHZ deformable shape dataset and a full episode from the television series Lost, and is shown to be superior to appearancebased approaches for matching non-rigid shape classes.
asian conference on computer vision | 2012
Ken Chatfield; Andrew Zisserman
This paper addresses the problem of object category retrieval in large unannotated image datasets. Our aim is to enable both fast learning of an object category model, and fast retrieval over the dataset. With these elements we show that new visual concepts can be learnt on-the-fly, given a text description, and so images of that category can then be retrieved from the dataset in realtime. To this end we compare state of the art encoding methods and introduce a novel cascade retrieval architecture, with a focus on achieving the best trade-off between three important performance measures for a realtime system of this kind, namely: (i) class accuracy, (ii) memory footprint, and (iii) speed. We show that an on-the-fly system is possible and compare its performance (using noisy training images) to that of using carefully curated images. For this evaluation we use the VOC 2007 dataset together with 100k images from ImageNet to act as distractors.
asian conference on computer vision | 2014
Ken Chatfield; Karen Simonyan; Andrew Zisserman
We investigate the gains in precision and speed, that can be obtained by using Convolutional Networks (ConvNets) for on-the-fly retrieval – where classifiers are learnt at run time for a textual query from downloaded images, and used to rank large image or video datasets.
international conference on multimedia retrieval | 2013
Kevin McGuinness; Noel E. O'Connor; Robin Aly; Franciska de Jong; Ken Chatfield; Omkar M. Parkhi; Relja Arandjelović; Andrew Zisserman; Matthijs Douze; Cordelia Schmid
We demonstrate a multimedia content information retrieval engine developed for audiovisual digital libraries targeted at media professionals. It is the first of three multimedia IR systems being developed by the AXES project. The system brings together traditional text IR and state-of-the-art content indexing and retrieval technologies to allow users to search and browse digital libraries in novel ways. Key features include: metadata and ASR search and filtering, on-the-fly visual concept classification (categories, faces, places, and logos), and similarity search (instances and faces).
International Journal of Multimedia Information Retrieval | 2015
Ken Chatfield; Relja Arandjelović; Omkar M. Parkhi; Andrew Zisserman
Proceedings TRECVid 2012 | 2012
Dan Oneata; Matthijs Douze; Jérôme Revaud; Schwenninger Jochen; Danila Potapov; Heng Wang; Zaid Harchaoui; Jakob J. Verbeek; Cordelia Schmid; Robin Aly; Kevin Mcguiness; Shu Chen; Noel E. O'Connor; Ken Chatfield; Omkar M. Parkhi; Relja Arandjelović; Andrew Zisserman; Fernando Basura; Tinne Tuytelaars
TRECVid 2013 | 2013
Robin Aly; Relja Arandjelović; Ken Chatfield; Matthijs Douze; Basura Fernando; Zaid Harchaoui; Kevin Mcguiness; Noel E. O'Connor; Dan Oneata; Omkar M. Parkhi; Danila Potapov; Jérôme Revaud; Cordelia Schmid; J.-L. Schwenninger; David Scott; Tinne Tuytelaars; Jakob J. Verbeek; Heng Wang; Andrew Zisserman
International Broadcasting Convention (IBC) 2014 Conference | 2014
T. Tommasi; Robin Aly; Kevin McGuinness; Ken Chatfield; Relja Arandjelović; Omkar M. Parkhi; Roeland Ordelman; Andrew Zisserman; Tinne Tuytelaars