Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Shimon Ullman is active.

Publication


Featured researches published by Shimon Ullman.


Human neurobiology | 1987

Shifts in selective visual attention: towards the underlying neural circuitry

Christof Koch; Shimon Ullman

Psychophysical and physiological evidence indicates that the visual system of primates and humans has evolved a specialized processing focus moving across the visual scene. This study addresses the question of how simple networks of neuron-like elements can account for a variety of phenomena associated with this shift of selective visual attention. Specifically, we propose the following: (1) A number of elementary features, such as color, orientation, direction of movement, disparity etc. are represented in parallel in different topographical maps, called the early representation. (2) There exists a selective mapping from the early topographic representation into a more central non-topographic representation, such that at any instant the central representation contains the properties of only a single location in the visual scene, the selected location. We suggest that this mapping is the principal expression of early selective visual attention. One function of selective attention is to fuse information from different maps into one coherent whole. (3) Certain selection rules determine which locations will be mapped into the central representation. The major rule, using the conspicuity of locations in the early representation, is implemented using a so-called Winner-Take-All network. Inhibiting the selected location in this network causes an automatic shift towards the next most conspicious location. Additional rules are proximity and similarity preferences. We discuss how these rules can be implemented in neuron-like networks and suggest a possible role for the extensive back-projection from the visual cortex to the LGN.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 1997

Face recognition: the problem of compensating for changes in illumination direction

Yael Adini; Yael Moses; Shimon Ullman

A face recognition system must recognize a face from a novel image despite the variations between images of the same face. A common approach to overcoming image variations because of changes in the illumination conditions is to use image representations that are relatively insensitive to these variations. Examples of such representations are edge maps, image intensity derivatives, and images convolved with 2D Gabor-like filters. Here we present an empirical study that evaluates the sensitivity of these representations to changes in illumination, as well as viewpoint and facial expression. Our findings indicated that none of the representations considered is sufficient by itself to overcome image variations because of a change in the direction of illumination. Similar results were obtained for changes due to viewpoint and expression. Image representations that emphasized the horizontal features were found to be less sensitive to changes in the direction of illumination. However, systems based only on such representations failed to recognize up to 20 percent of the faces in our database. Humans performed considerably better under the same conditions. We discuss possible reasons for this superiority and alternative methods for overcoming illumination effects in recognition.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 1991

Recognition by linear combinations of models

Shimon Ullman; Ronen Basri

An approach to visual object recognition in which a 3D object is represented by the linear combination of 2D images of the object is proposed. It is shown that for objects with sharp edges as well ...An approach to visual object recognition in which a 3D object is represented by the linear combination of 2D images of the object is proposed. It is shown that for objects with sharp edges as well as with smooth bounding contours, the set of possible images of a given object is embedded in a linear space spanned by a small number of views. For objects with sharp edges, the linear combination representation is exact. For objects with smooth boundaries, it is an approximation that often holds over a wide range of viewing angles. Rigid transformations (with or without scaling) can be distinguished from more general linear transformations of the object by testing certain constraints placed on the coefficients of the linear combinations. Three alternative methods of determining the transformation that matches a model to a given image are proposed. >


Cognition | 1989

Aligning pictorial descriptions: An approach to object recognition

Shimon Ullman

This paper examines the problem of shape-based object recognition, and proposes a new approach, the alignment of pictorial descriptions. The first part of the paper reviews general approaches to visual object recognition, and divides these approaches into three broad classes: invariant properties methods, object decomposition methods, and alignment methods. The second part presents the alignment method. In this approach the recognition process is divided into two stages. The first determines the transformation in space that is necessary to bring the viewed object into alignment with possible object models. This stage can proceed on the basis of minimal information, such as the objects dominant orientation, or a small number of corresponding feature points in the object and model. The second stage determines the model that best matches the viewed object. At this stage, the search is over all the possible object models, but not over their possible views, since the transformation has already been determined uniquely in the alignment stage. The proposed alignment method also uses abstract description, but unlike structural description methods it uses them pictorially, rather than in symbolic structural descriptions.


Trends in Neurosciences | 1983

The Measurement of Visual Motion

Ellen C. Hildreth; Shimon Ullman

Abstract Visual motion provides useful information about the surrounding environment, which biological visual systems have evolved to extract and utilize. The first problem in analysing visual motion is the measurement of motion; this has proved to be surprisingly difficult. The human visual system appears to solve it efficiently using a combination of at least two different methods. These methods are discussed, together with some unsolved problems and their possible implications for neurophysiological studies.


Nature Neuroscience | 2002

Visual features of intermediate complexity and their use in classification

Shimon Ullman; Michel Vidal-Naquet; Erez Sali

The human visual system analyzes shapes and objects in a series of stages in which stimulus features of increasing complexity are extracted and analyzed. The first stages use simple local features, and the image is subsequently represented in terms of larger and more complex features. These include features of intermediate complexity and partial object views. The nature and use of these higher-order representations remains an open question in the study of visual processing by the primate cortex. Here we show that intermediate complexity (IC) features are optimal for the basic visual task of classification. Moderately complex features are more informative for classification than very simple or very complex ones, and so they emerge naturally by the simple coding principle of information maximization with respect to a class of images. Our findings suggest a specific role for IC features in visual processing and a principle for their extraction.


International Journal of Computer Vision | 1990

Recognizing solid objects by alignment with an image

Daniel P. Huttenlocher; Shimon Ullman

In this paper we consider the problem of recognizing solid objects from a single two-dimensional image of a three-dimensional scene. We develop a new method for computing a transformation from a three-dimensional model coordinate frame to the two-dimensional image coordinate frame, using three pairs of model and image points. We show that this transformation always exists for three noncollinear points, and is unique up to a reflective ambiguity. The solution method is closed-form and only involves second-order equations. We have implemented a recognition system that uses this transformation method to determine possible alignments of a model with an image. Each of these hypothesized matches is verified by comparing the entire edge contours of the aligned object with the image edges. Using the entire edge contours for verification, rather than a few local feature points, reduces the chance of finding false matches. The system has been tested on partly occluded objects in highly cluttered scenes.


Proceedings of the Royal Society of London. Series B, Biological sciences | 1979

The Interpretation of Structure from Motion

Shimon Ullman

The interpretation of structure from motion is examined from a computional point of view. The question addressed is how the three dimensional structure and motion of objects can be inferred from the two dimensional transformations of their projected images when no three dimensional information is conveyed by the individual projections. The following scheme is proposed: (i) divide the image into groups of four elements each; (ii) test each group for a rigid interpretation; (iii) combine the results obtained in (ii). It is shown that this scheme will correctly decompose scenes containing arbitrary rigid objects in motion, recovering their three dimensional structure and motion. The analysis is based primarily on the ʻstructure from motion’ theorem which states that the structure of four non-coplanar points is recoverable from three orthographic projections. The interpretation scheme is extended to cover perspective projections, and its psychological relevance is discussed.


Behavioral and Brain Sciences | 1980

Against Direct Perception

Shimon Ullman

Central to contemporary cognitive science is the notion that mental processes involve computations defined over internal representations. This view stands in sharp contrast to the “direct approach” to visual perception and cognition, whose most prominent proponent has been J.J. Gibson. In the direct theory, perception does not involve computations of any sort; it is the result of the direct pickup of available information. The publication of Gibsons recent book (Gibson 1979) offers an opportunity to examine his approach, and, more generally, to contrast the theory of direct perception with the computational/representational view. In the first part of the present article (Sections 2–3) the notion of “direct perception” is examined from a theoretical standpoint, and a number of objections are raised. Section 4 is a “case study”: the problem of perceiving the three-dimensional shape of moving objects is examined. This problem, which has been extensively studied within the immediate perception framework, serves to illustrate some of the inherent shortcomings of that approach. Finally, in Section 5, an attempt is made to place the theory of direct perception in perspective by embedding it in a more comprehensive framework.


european conference on computer vision | 2002

Class-Specific, Top-Down Segmentation

Eran Borenstein; Shimon Ullman

In this paper we present a novel class-based segmentation method, which is guided by a stored representation of the shape of objects within a general class (such as horse images). The approach is different from bottom-up segmentation methods that primarily use the continuity of grey-level, texture, and bounding contours. We show that the method leads to markedly improved segmentation results and can deal with significant variation in shape and varying backgrounds. We discuss the relative merits of class-specific and general image-based segmentation methods and suggest how they can be usefully combined.

Collaboration


Dive into the Shimon Ullman's collaboration.

Top Co-Authors

Avatar

Daniel Harari

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Boris Epshtein

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Erez Sali

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Evgeniy Bart

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Yael Moses

Interdisciplinary Center Herzliya

View shared research outputs
Top Co-Authors

Avatar

Dror Zur

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Guy Ben-Yosef

Ben-Gurion University of the Negev

View shared research outputs
Top Co-Authors

Avatar

Liav Assif

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Michel Vidal-Naquet

Weizmann Institute of Science

View shared research outputs
Top Co-Authors

Avatar

Joshua B. Tenenbaum

Massachusetts Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge