Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Catherine Wah is active.

Publication


Featured researches published by Catherine Wah.


european conference on computer vision | 2010

Visual recognition with humans in the loop

Steve Branson; Catherine Wah; Florian Schroff; Boris Babenko; Peter Welinder; Pietro Perona; Serge J. Belongie

We present an interactive, hybrid human-computer method for object classification. The method applies to classes of objects that are recognizable by people with appropriate expertise (e.g., animal species or airplane model), but not (in general) by people without such expertise. It can be seen as a visual version of the 20 questions game, where questions based on simple visual attributes are posed interactively. The goal is to identify the true class while minimizing the number of questions asked, using the visual content of the image. We introduce a general framework for incorporating almost any off-the-shelf multi-class object recognition algorithm into the visual 20 questions game, and provide methodologies to account for imperfect user responses and unreliable computer vision algorithms. We evaluate our methods on Birds-200, a difficult dataset of 200 tightly-related bird species, and on the Animals With Attributes dataset. Our results demonstrate that incorporating user input drives up recognition accuracy to levels that are good enough for practical applications, while at the same time, computer vision reduces the amount of human interaction required.


international conference on computer vision | 2011

Multiclass recognition and part localization with humans in the loop

Catherine Wah; Steve Branson; Pietro Perona; Serge J. Belongie

We propose a visual recognition system that is designed for fine-grained visual categorization. The system is composed of a machine and a human user. The user, who is unable to carry out the recognition task by himself, is interactively asked to provide two heterogeneous forms of information: clicking on object parts and answering binary questions. The machine intelligently selects the most informative question to pose to the user in order to identify the objects class as quickly as possible. By leveraging computer vision and analyzing the user responses, the overall amount of human effort required, measured in seconds, is minimized. We demonstrate promising results on a challenging dataset of uncropped images, achieving a significant average reduction in human effort over previous methods.


computer vision and pattern recognition | 2013

Style Finder: Fine-Grained Clothing Style Detection and Retrieval

Wei Di; Catherine Wah; Anurag Bhardwaj; Robinson Piramuthu; Neel Sundaresan

With the rapid proliferation of smartphones and tablet computers, search has moved beyond text to other modalities like images and voice. For many applications like Fashion, visual search offers a compelling interface that can capture stylistic visual elements beyond color and pattern that cannot be as easily described using text. However, extracting and matching such attributes remains an extremely challenging task due to high variability and deformability of clothing items. In this paper, we propose a fine-grained learning model and multimedia retrieval framework to address this problem. First, an attribute vocabulary is constructed using human annotations obtained on a novel fine-grained clothing dataset. This vocabulary is then used to train a fine-grained visual recognition system for clothing styles. We report benchmark recognition and retrieval results on Womens Fashion Coat Dataset and illustrate potential mobile applications for attribute-based multimedia retrieval of clothing items and image annotation.


International Journal of Computer Vision | 2014

The Ignorant Led by the Blind: A Hybrid Human---Machine Vision System for Fine-Grained Categorization

Steve Branson; Grant Van Horn; Catherine Wah; Pietro Perona; Serge J. Belongie

We present a visual recognition system for fine-grained visual categorization. The system is composed of a human and a machine working together and combines the complementary strengths of computer vision algorithms and (non-expert) human users. The human users provide two heterogeneous forms of information object part clicks and answers to multiple choice questions. The machine intelligently selects the most informative question to pose to the user in order to identify the object class as quickly as possible. By leveraging computer vision and analyzing the user responses, the overall amount of human effort required, measured in seconds, is minimized. Our formalism shows how to incorporate many different types of computer vision algorithms into a human-in-the-loop framework, including standard multiclass methods, part-based methods, and localized multiclass and attribute methods. We explore our ideas by building a field guide for bird identification. The experimental results demonstrate the strength of combining ignorant humans with poor-sighted machines the hybrid system achieves quick and accurate bird identification on a dataset containing 200 bird species.


computer vision and pattern recognition | 2014

Similarity Comparisons for Interactive Fine-Grained Categorization

Catherine Wah; Grant Van Horn; Steve Branson; Subhransu Maji; Pietro Perona; Serge J. Belongie

Current human-in-the-loop fine-grained visual categorization systems depend on a predefined vocabulary of attributes and parts, usually determined by experts. In this work, we move away from that expert-driven and attribute-centric paradigm and present a novel interactive classification system that incorporates computer vision and perceptual similarity metrics in a unified framework. At test time, users are asked to judge relative similarity between a query image and various sets of images, these general queries do not require expert-defined terminology and are applicable to other domains and basic-level categories, enabling a flexible, efficient, and scalable system for fine-grained categorization with humans in the loop. Our system outperforms existing state-of-the-art systems for relevance feedback-based image retrieval as well as interactive classification, resulting in a reduction of up to 43% in the average number of questions needed to correctly classify an image.


workshop on applications of computer vision | 2015

Learning Localized Perceptual Similarity Metrics for Interactive Categorization

Catherine Wah; Subhransu Maji; Serge J. Belongie

Current similarity-based approaches to interactive fine grained categorization rely on learning metrics from holistic perceptual measurements of similarity between objects or images. However, making a single judgment of similarity at the object level can be a difficult or overwhelming task for the human user to perform. Secondly, a single general metric of similarity may not be able to adequately capture the minute differences that discriminate fine-grained categories. In this work, we propose a novel approach to interactive categorization that leverages multiple perceptual similarity metrics learned from localized and roughly aligned regions across images, reporting state-of-the-art results and outperforming methods that use a single nonlocalized similarity metric.


Advances in Water Resources | 2011

The Caltech-UCSD Birds-200-2011 Dataset

Catherine Wah; Steve Branson; Peter Welinder; Pietro Perona; Serge J. Belongie


Archive | 2010

Caltech-UCSD Birds 200

Peter Welinder; Steve Branson; Takeshi Mita; Catherine Wah; Florian Schroff; Serge J. Belongie; Pietro Perona


computer vision and pattern recognition | 2013

Attribute-Based Detection of Unfamiliar Classes with Humans in the Loop

Catherine Wah; Serge J. Belongie


Archive | 2011

Crowdsourcing and Its Applications in Computer Vision

Catherine Wah

Collaboration


Dive into the Catherine Wah's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Pietro Perona

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Steve Branson

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Peter Welinder

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Grant Van Horn

California Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Subhransu Maji

University of Massachusetts Amherst

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Boris Babenko

University of California

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge