Andrea Frome
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Andrea Frome.
european conference on computer vision | 2004
Andrea Frome; Daniel Huber; Ravi Krishna Kolluri; Thomas Bülow; Jitendra Malik
Recognition of three dimensional (3D) objects in noisy and cluttered scenes is a challenging problem in 3D computer vision. One approach that has been successful in past research is the regional shape descriptor. In this paper, we introduce two new regional shape descriptors: 3D shape contexts and harmonic shape contexts. We evaluate the performance of these descriptors on the task of recognizing vehicles in range scans of scenes using a database of 56 cars. We compare the two novel descriptors to an existing descriptor, the spin image, showing that the shape context based descriptors have a higher recognition rate on noisy scenes and that 3D shape contexts outperform the others on cluttered scenes.
international conference on computer vision | 2007
Andrea Frome; Yoram Singer; Fei Sha; Jitendra Malik
We address the problem of visual category recognition by learning an image-to-image distance function that attempts to satisfy the following property: the distance between images from the same category should be less than the distance between images from different categories. We use patch-based feature vectors common in object recognition work as a basis for our image-to-image distance functions. Our large-margin formulation for learning the distance functions is similar to formulations used in the machine learning literature on distance metric learning, however we differ in that we learn local distance functions¿a different parameterized function for every image of our training set¿whereas typically a single global distance function is learned. This was a novel approach first introduced in Frome, Singer, & Malik, NIPS 2006. In that work we learned the local distance functions independently, and the outputs of these functions could not be compared at test time without the use of additional heuristics or training. Here we introduce a different approach that has the advantage that it learns distance functions that are globally consistent in that they can be directly compared for purposes of retrieval and classification. The output of the learning algorithm are weights assigned to the image features, which is intuitively appealing in the computer vision setting: some features are more salient than others, and which are more salient depends on the category, or image, being considered. We train and test using the Caltech 101 object recognition benchmark.
european conference on computer vision | 2014
Jia Deng; Nan Ding; Yangqing Jia; Andrea Frome; Kevin P. Murphy; Samy Bengio; Yuan Li; Hartmut Neven; Hartwig Adam
In this paper we study how to perform object classification in a principled way that exploits the rich structure of real world labels. We develop a new model that allows encoding of flexible relations between labels. We introduce Hierarchy and Exclusion (HEX) graphs, a new formalism that captures semantic relations between any two labels applied to the same object: mutual exclusion, overlap and subsumption. We then provide rigorous theoretical analysis that illustrates properties of HEX graphs such as consistency, equivalence, and computational implications of the graph structure. Next, we propose a probabilistic classification model based on HEX graphs and show that it enjoys a number of desirable properties. Finally, we evaluate our method using a large-scale benchmark. Empirical results demonstrate that our model can significantly improve object classification by exploiting the label relations.
international conference on computer vision | 2009
Andrea Frome; German Cheung; Ahmad Abdulkader; Marco Zennaro; Bo Wu; Alessandro Bissacco; Hartwig Adam; Hartmut Neven; Luc Vincent
The last two years have witnessed the introduction and rapid expansion of products based upon large, systematically-gathered, street-level image collections, such as Google Street View, EveryScape, and Mapjack. In the process of gathering images of public spaces, these projects also capture license plates, faces, and other information considered sensitive from a privacy standpoint. In this work, we present a system that addresses the challenge of automatically detecting and blurring faces and license plates for the purpose of privacy protection in Google Street View. Though some in the field would claim face detection is “solved”, we show that state-of-the-art face detectors alone are not sufficient to achieve the recall desired for large-scale privacy protection. In this paper we present a system that combines a standard sliding-window detector tuned for a high recall, low-precision operating point with a fast post-processing stage that is able to remove additional false positives by incorporating domain-specific information not available to the sliding-window detector. Using a completely automatic system, we are able to sufficiently blur more than 89% of faces and 94 – 96% of license plates in evaluation sets sampled from Google Street View imagery.
neural information processing systems | 2013
Andrea Frome; Gregory S. Corrado; Jonathon Shlens; Samy Bengio; Jeffrey Dean; Marc'Aurelio Ranzato; Tomas Mikolov
international conference on learning representations | 2014
Mohammad Norouzi; Tomas Mikolov; Samy Bengio; Yoram Singer; Jonathon Shlens; Andrea Frome; Greg Corrado; Jeffrey Dean
neural information processing systems | 2006
Andrea Frome; Yoram Singer; Jitendra Malik
Archive | 2009
Sergey Ioffe; Lance Williams; Dennis Strelow; Andrea Frome; Luc Vincent
Archive | 2010
Charles Rosenberg; Jingbin Wang; Sarah Moussa; Erik Murphy-Chutorian; Andrea Frome; Yoram Singer; Radhika Malpani
Archive | 2012
Andrea Frome; German Cheung; Ahmad Abdulkader; Marco Zennaro; Bo Wu; Alessandro Bissallo; Hartmut Neven; Luc Vincent; Hartwig Adam