Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Alex Holub is active.

Publication


Featured researches published by Alex Holub.


computer vision and pattern recognition | 2008

Entropy-based active learning for object recognition

Alex Holub; Pietro Perona; Michael C. Burl

Most methods for learning object categories require large amounts of labeled training data. However, obtaining such data can be a difficult and time-consuming endeavor. We have developed a novel, entropy-based ldquoactive learningrdquo approach which makes significant progress towards this problem. The main idea is to sequentially acquire labeled data by presenting an oracle (the user) with unlabeled images that will be particularly informative when labeled. Active learning adaptively prioritizes the order in which the training examples are acquired, which, as shown by our experiments, can significantly reduce the overall number of training examples required to reach near-optimal performance. At first glance this may seem counter-intuitive: how can the algorithm know whether a group of unlabeled images will be informative, when, by definition, there is no label directly associated with any of the images? Our approach is based on choosing an image to label that maximizes the expected amount of information we gain about the set of unlabeled images. The technique is demonstrated in several contexts, including improving the efficiency of Web image-search queries and open-world visual learning by an autonomous agent. Experiments on a large set of 140 visual object categories taken directly from text-based Web image searches show that our technique can provide large improvements (up to 10 x reduction in the number of training examples needed) over baseline techniques.


computer vision and pattern recognition | 2005

A discriminative framework for modelling object classes

Alex Holub; Pietro Perona

Here we explore a discriminative learning method on underlying generative models for the purpose of discriminating between object categories. Visual recognition algorithms learn models from a set of training examples. Generative models learn their representations by considering data from a single class. Generative models are popular in computer vision for many reasons, including their ability to elegantly incorporate prior knowledge and to handle correspondences between object parts and detected features. However, generative models are often inferior to discriminative models during classification tasks. We study a discriminative approach to learning object categories which maintains the representational power of generative learning, but trains the generative models in a discriminative manner. The discriminatively trained models perform better during classification tasks as a result of selecting discriminative sets of features. We conclude by proposing a multi-class object recognition system which initially trains object classes in a generative manner, identifies subsets of similar classes with high confusion, and finally trains models for these subsets in a discriminative manner to realize gains in classification performance.


Neural Networks | 2005

2005 Special Issue: Bayesian approach to feature selection and parameter tuning for support vector machine classifiers

Carl Gold; Alex Holub; Peter Sollich

A Bayesian point of view of SVM classifiers allows the definition of a quantity analogous to the evidence in probabilistic models. By maximizing this one can systematically tune hyperparameters and, via automatic relevance determination (ARD), select relevant input features. Evidence gradients are expressed as averages over the associated posterior and can be approximated using Hybrid Monte Carlo (HMC) sampling. We describe how a Nyström approximation of the Gram matrix can be used to speed up sampling times significantly while maintaining almost unchanged classification accuracy. In experiments on classification problems with a significant number of irrelevant features this approach to ARD can give a significant improvement in classification performance over more traditional, non-ARD, SVM systems. The final tuned hyperparameter values provide a useful criterion for pruning irrelevant features, and we define a measure of relevance with which to determine systematically how many features should be removed. This use of ARD for hard feature selection can improve classification accuracy in non-ARD SVMs. In the majority of cases, however, we find that in data sets constructed by human domain experts the performance of non-ARD SVMs is largely insensitive to the presence of some less relevant features. Eliminating such features via ARD then does not improve classification accuracy, but leads to impressive reductions in the number of features required, by up to 75%.


ieee international conference on automatic face & gesture recognition | 2008

Unsupervised clustering for google searches of celebrity images

Alex Holub; Pierre Moreels; Pietro Perona

How do we identify images of the same person in photo albums? How can we find images of a particular celebrity using Web image search engines? These types of tasks require solving numerous challenging issues in computer vision including: detecting whether an image contains a face, maintaining robustness to lighting, pose, occlusion, scale, and image quality, and using appropriate distance metrics to identify and compare different faces. In this paper we present a complete system which yields good performance on challenging tasks involving face recognition including image retrieval, unsupervised clustering of faces, and increasing precision of dasiaGoogle Imagepsila searches. All tasks use highly variable real data obtained from raw image searches on the web.


computer vision and pattern recognition | 2007

On Constructing Facial Similarity Maps

Alex Holub; Yun-hsueh Liu; Pietro Perona

Automatically determining facial similarity is a difficult and open question in computer vision. The problem is complicated both because it is unclear what facial features humans use to determine facial similarity and because facial similarity is subjective in nature: similarity judgements change from person to person. In this work we suggest a system which places facial similarity on a solid computational footing. First we describe methods for acquiring facial similarity ratings from humans in an efficient manner. Next we show how to create feature vector representations for each face by extracted patches around facial key-points. Finally we show how to use the acquired similarity ratings to learn functional mapping which project facial-feature vectors into face spaces which correspond to our notions of facial similarity. We use different collections of images to both create and validate the face spaces including: perceptual similarity data obtained from humans, morphed faces between two different individuals, and the CMU PIE collection which contains images of the same individual under different lighting conditions. We demonstrate that using our methods we can effectively create face spaces which correspond to human notions of facial similarity.


Journal of Comparative Physiology A-neuroethology Sensory Neural and Behavioral Physiology | 2001

Temporal population code of concurrent vocal signals in the auditory midbrain.

Deana A. Bodnar; Alex Holub; Bruce R. Land; Joseph F. Skovira; Andrew H. Bass

Abstract. Unique patterns of spike activity across neuron populations have been implicated in the coding of complex sensory stimuli. Delineating the patterns of neural activity in response to varying stimulus parameters and their relationships to the tuning characteristics of individual neurons is essential to ascertaining the nature of population coding within the brain. Here, we address these points in the midbrain coding of concurrent vocal signals of a sound-producing fish, the plainfin midshipman. Midshipman produce multiharmonic vocalizations which frequently overlap to produce beats. We used multivariate statistical analysis from single-unit recordings across multiple animals to assess the presence of a temporal population code. Our results show that distinct patterns of temporal activity emerge among midbrain neurons in response to concurrent signals that vary in their difference frequency. These patterns can serve to code beat difference frequencies. The patterns directly result from the differential temporal coding of difference frequency by individual neurons. Difference frequency encoding, based on temporal patterns of activity, could permit the segregation of concurrent vocal signals on time scales shorter than codes requiring averaging. Given the ubiquity across vertebrates of auditory midbrain tuning to the temporal structure of acoustic signals, a similar temporal population code is likely present in other species.


computer vision and pattern recognition | 2009

Towards unlocking web video: Automatic people tracking and clustering

Alex Holub; Pierre Moreels; Atiq Islam; Andrei Peter Makhanov; Rui Yang

This paper describes a system for automatically extracting meta-information on people from videos on the Web. The system contains multiple modules which automatically track people, including both faces and bodies, and clusters the people into distinct groups. We present new technology and significantly modify existing algorithms for body-detection, shot-detection and grouping, tracking, and track-clustering within our system. The system was designed to work effectivity on Web content, and thus exhibits robust tracking and clustering behavior over a broad spectrum of professional and semi-professional video content. In order to quantify and evaluate our system we created a large ground-truth data-set of people within video. Finally, we provide actual video examples of our algorithm and find that the results are quite strong over a broad range of content.


Archive | 2007

Caltech-256 Object Category Dataset

Gregory Griffin; Alex Holub; Pietro Perona


Archive | 2006

Using relevance feedback in face recognition

Keren Perlmutter; Sharon M. Perlmutter; Joshua Alspector; Mark Everingham; Alex Holub; Andrew Zisserman; Pietro Perona


international conference on computer vision | 2005

Combining generative models and Fisher kernels for object recognition

Alex Holub; Max Welling; Pietro Perona

Collaboration


Dive into the Alex Holub's collaboration.

Top Co-Authors

Avatar

Pietro Perona

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Pierre Moreels

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Max Welling

University of Amsterdam

View shared research outputs
Top Co-Authors

Avatar

Carl Gold

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Dennis D.M. O'Leary

Salk Institute for Biological Studies

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge