Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Krista A. Ehinger is active.

Publication


Featured researches published by Krista A. Ehinger.


computer vision and pattern recognition | 2010

SUN database: Large-scale scene recognition from abbey to zoo

Jianxiong Xiao; James Hays; Krista A. Ehinger; Aude Oliva; Antonio Torralba

Scene categorization is a fundamental problem in computer vision. However, scene understanding research has been constrained by the limited scope of currently-used databases which do not capture the full variety of scene categories. Whereas standard databases for object categorization contain hundreds of different classes of objects, the largest available dataset of scene categories contains only 15 classes. In this paper we propose the extensive Scene UNderstanding (SUN) database that contains 899 categories and 130,519 images. We use 397 well-sampled categories to evaluate numerous state-of-the-art algorithms for scene recognition and establish new bounds of performance. We measure human scene classification performance on the SUN database and compare this with computational methods. Additionally, we study a finer-grained scene representation to detect scenes embedded inside of larger scenes.


international conference on computer vision | 2009

Learning to predict where humans look

Tilke Judd; Krista A. Ehinger; Antonio Torralba

For many applications in graphics, design, and human computer interaction, it is essential to understand where humans look in a scene. Where eye tracking devices are not a viable option, models of saliency can be used to predict fixation locations. Most saliency approaches are based on bottom-up computation that does not consider top-down image semantics and often does not match actual eye movements. To address this problem, we collected eye tracking data of 15 viewers on 1003 images and use this database as training and testing examples to learn a model of saliency based on low, middle and high-level image features. This large database of eye tracking data is publicly available with this paper.


International Journal of Computer Vision | 2016

SUN Database: Exploring a Large Collection of Scene Categories

Jianxiong Xiao; Krista A. Ehinger; James Hays; Antonio Torralba; Aude Oliva

Progress in scene understanding requires reasoning about the rich and diverse visual environments that make up our daily experience. To this end, we propose the Scene Understanding database, a nearly exhaustive collection of scenes categorized at the same level of specificity as human discourse. The database contains 908 distinct scene categories and 131,072 images. Given this data with both scene and object labels available, we perform in-depth analysis of co-occurrence statistics and the contextual relationship. To better understand this large scale taxonomy of scene categories, we perform two human experiments: we quantify human scene recognition accuracy, and we measure how typical each image is of its assigned scene category. Next, we perform computational experiments: scene recognition with global image features, indoor versus outdoor classification, and “scene detection,” in which we relax the assumption that one image depicts only one scene category. Finally, we relate human experiments to machine performance and explore the relationship between human and machine recognition errors and the relationship between image “typicality” and machine recognition accuracy.


computer vision and pattern recognition | 2012

Recognizing scene viewpoint using panoramic place representation

Jianxiong Xiao; Krista A. Ehinger; Aude Oliva; Antonio Torralba

We introduce the problem of scene viewpoint recognition, the goal of which is to classify the type of place shown in a photo, and also recognize the observers viewpoint within that category of place. We construct a database of 360° panoramic images organized into 26 place categories. For each category, our algorithm automatically aligns the panoramas to build a full-view representation of the surrounding place. We also study the symmetry properties and canonical viewpoint of each place category. At test time, given a photo of a scene, the model can recognize the place category, produce a compass-like indication of the observers most likely viewpoint within that place, and use this information to extrapolate beyond the available view, filling in the probable visual layout that would appear beyond the boundary of the photo.


Frontiers in Psychology | 2012

Rethinking the Role of Top-Down Attention in Vision: Effects Attributable to a Lossy Representation in Peripheral Vision

Ruth Rosenholtz; Jie Huang; Krista A. Ehinger

According to common wisdom in the field of visual perception, top-down selective attention is required in order to bind features into objects. In this view, even simple tasks, such as distinguishing a rotated T from a rotated L, require selective attention since they require feature binding. Selective attention, in turn, is commonly conceived as involving volition, intention, and at least implicitly, awareness. There is something non-intuitive about the notion that we might need so expensive (and possibly human) a resource as conscious awareness in order to perform so basic a function as perception. In fact, we can carry out complex sensorimotor tasks, seemingly in the near absence of awareness or volitional shifts of attention (“zombie behaviors”). More generally, the tight association between attention and awareness, and the presumed role of attention on perception, is problematic. We propose that under normal viewing conditions, the main processes of feature binding and perception proceed largely independently of top-down selective attention. Recent work suggests that there is a significant loss of information in early stages of visual processing, especially in the periphery. In particular, our texture tiling model (TTM) represents images in terms of a fixed set of “texture” statistics computed over local pooling regions that tile the visual input. We argue that this lossy representation produces the perceptual ambiguities that have previously been as ascribed to a lack of feature binding in the absence of selective attention. At the same time, the TTM representation is sufficiently rich to explain performance in such complex tasks as scene gist recognition, pop-out target search, and navigation. A number of phenomena that have previously been explained in terms of voluntary attention can be explained more parsimoniously with the TTM. In this model, peripheral vision introduces a specific kind of information loss, and the information available to an observer varies greatly depending upon shifts of the point of gaze (which usually occur without awareness). The available information, in turn, provides a key determinant of the visual system’s capabilities and deficiencies. This scheme dissociates basic perceptual operations, such as feature binding, from both top-down attention and conscious awareness.


Attention Perception & Psychophysics | 2008

The role of color in visual search in real-world scenes: Evidence from contextual cuing

Krista A. Ehinger; James R. Brockmole

Because the importance of color in visual tasks such as object identification and scene memory has been debated, we sought to determine whether color is used to guide visual search in contextual cuing with real-world scenes. In Experiment 1, participants searched for targets in repeated scenes that were shown in one of three conditions: natural colors, unnatural colors that remained consistent across repetitions, and unnatural colors that changed on every repetition. We found that the pattern of learning was the same in all three conditions. In Experiment 2, we did a transfer test in which the repeating scenes were shown in consistent colors that suddenly changed on the last block of the experiment. The color change had no effect on search times, relative to a condition in which the colors did not change. In Experiments 3 and 4, we replicated Experiments 1 and 2, using scenes from a color-diagnostic category of scenes, and obtained similar results. We conclude that color is not used to guide visual search in real-world contextual cuing, a finding that constrains the role of color in scene identification and recognition processes.


Frontiers in Psychology | 2013

Basic level scene understanding: categories, attributes and structures

Jianxiong Xiao; James Hays; Bryan C. Russell; Genevieve Patterson; Krista A. Ehinger; Antonio Torralba; Aude Oliva

A longstanding goal of computer vision is to build a system that can automatically understand a 3D scene from a single image. This requires extracting semantic concepts and 3D information from 2D images which can depict an enormous variety of environments that comprise our visual world. This paper summarizes our recent efforts toward these goals. First, we describe the richly annotated SUN database which is a collection of annotated images spanning 908 different scene categories with object, attribute, and geometric labels for many scenes. This database allows us to systematically study the space of scenes and to establish a benchmark for scene and object recognition. We augment the categorical SUN database with 102 scene attributes for every image and explore attribute recognition. Finally, we present an integrated system to extract the 3D structure of the scene and objects depicted in an image.


Psychonomic Bulletin & Review | 2015

Through the looking-glass: Objects in the mirror are less real.

Preeti Sareen; Krista A. Ehinger; Jeremy M. Wolfe

Is an object reflected in a mirror perceived differently from an object that is seen directly? We asked observers to label “everything” in photographs of real-world scenes. Some scenes contained a mirror in which objects could be seen. Reflected objects received significantly fewer labels than did their nonreflected counterparts. If an object was visible only as a reflection, it was labeled more often than a reflected object that appeared both as a reflection and nonreflected in the room. These unique reflected objects were still not labeled more often than the unique nonreflected objects in the room. In a second experiment, we used a change blindness paradigm in which equivalent object changes occurred in the nonreflected and reflected parts of the scene. Reaction times were longer and accuracy was lower for finding the changes in reflections. These results suggest that reflected information is easily discounted when processing images of natural scenes.


Pattern Recognition | 2017

A novel graph-based optimization framework for salient object detection

Jinxia Zhang; Krista A. Ehinger; Haikun Wei; Kanjian Zhang; Jingyu Yang

In traditional graph-based optimization framework for salient object detection, an image is over-segmented into superpixels and mapped to one single graph. The saliency value of each superpixel is then computed based on the similarity between connected nodes and the saliency related queries. When applying the traditional graph-based optimization framework to the salient object detection problem in natural scene images, we observe at least two limitations: only one graph is employed to describe the information contained in an image and no cognitive property about visual saliency is explicitly modeled in the optimization framework. In this work, we propose a novel graph-based optimization framework for salient object detection. Firstly, we employ multiple graphs in our optimization framework. A natural scene image is usually complex, employing multiple graphs from different image properties can better describe the complex information contained in the image. Secondly, we model one popular cognitive property about visual saliency (visual rarity) in our graph-based optimization framework, making this framework more suitable for saliency detection problem. Specifically, we add a regularization term to constrain the saliency value of each superpixel according to visual rarity in our optimization framework. Our experimental results on four benchmark databases with comparisons to fifteen representative methods demonstrate that our graph-based optimization framework is effective and computationally efficient. HighlightsA novel graph-based optimization framework for salient object detection is proposed in the paper.Multiple graphs are employed in our optimization framework to better describe a natural scene image.Visual rarity is modeled as a regularization term in our framework to better detect saliency.Experimental results on four datasets with fifteen methods prove the effectiveness of our method.


international conference on computer graphics and interactive techniques | 2012

Basic level scene understanding: from labels to structure and beyond

Jianxiong Xiao; Bryan C. Russell; James Hays; Krista A. Ehinger; Aude Oliva; Antonio Torralba

An early goal of computer vision was to build a system that could automatically understand a 3D scene just by looking. This requires not only the ability to extract 3D information from image information alone, but also to handle the large variety of different environments that comprise our visual world. This paper summarizes our recent efforts toward these goals. First, we describe the SUN database, which is a collection of annotated images spanning 908 different scene categories. This database allows us to systematically study the space of possible everyday scenes and to establish a benchmark for scene and object recognition. We also explore ways of coping with the variety of viewpoints within these scenes. For this, we have introduced a database of 360° panoramic images for many of the scene categories in the SUN database and have explored viewpoint recognition within the environments. Finally, we describe steps toward a unified 3D parsing of everyday scenes: (i) the ability to localize geometric primitives in images, such as cuboids and cylinders, which often comprise many everyday objects, and (ii) an integrated system to extract the 3D structure of the scene and objects depicted in an image.

Collaboration


Dive into the Krista A. Ehinger's collaboration.

Top Co-Authors

Avatar

Aude Oliva

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jeremy M. Wolfe

Brigham and Women's Hospital

View shared research outputs
Top Co-Authors

Avatar

Antonio Torralba

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

James Hays

Georgia Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Ruth Rosenholtz

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Erich W. Graf

University of Southampton

View shared research outputs
Top Co-Authors

Avatar

Wendy J. Adams

University of Southampton

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge