Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Dimitris Samaras is active.

Publication


Featured researches published by Dimitris Samaras.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2006

Face recognition from a single training image under arbitrary unknown lighting using spherical harmonics

Lei Zhang; Dimitris Samaras

In this paper, we propose two novel methods for face recognition under arbitrary unknown lighting by using spherical harmonics illumination representation, which require only one training image per subject and no 3D shape information. Our methods are based on the result which demonstrated that the set of images of a convex Lambertian object obtained under a wide variety of lighting conditions can be approximated accurately by a low-dimensional linear subspace. We provide two methods to estimate the spherical harmonic basis images spanning this space from just one image. Our first method builds the statistical model based on a collection of 2D basis images. We demonstrate that, by using the learned statistics, we can estimate the spherical harmonic basis images from just one image taken under arbitrary illumination conditions if there is no pose variation. Compared to the first method, the second method builds the statistical models directly in 3D spaces by combining the spherical harmonic illumination representation and a 3D morphable model of human faces to recover basis images from images across both poses and illuminations. After estimating the basis images, we use the same recognition scheme for both methods: we recognize the face for which there exists a weighted combination of basis images that is the closest to the test face image. We provide a series of experiments that achieve high recognition rates, under a wide range of illumination conditions, including multiple sources of illumination. Our methods achieve comparable levels of accuracy with methods that have much more onerous training data requirements. Comparison of the two methods is also provided.


computer vision and pattern recognition | 2003

Face recognition under variable lighting using harmonic image exemplars

Lei Zhang; Dimitris Samaras

We propose a new approach for face recognition under arbitrary illumination conditions, which requires only one training image per subject (if there is no pose variation) and no 3D shape information. Our method is based on the result of Basri and Jacobs (2001), which demonstrated that the set of images of a convex Lambertian object obtained under a wide variety of lighting conditions can be approximated accurately by a low-dimensional linear subspace. In this paper, we show that we can recover basis images spanning this space from just one image taken under arbitrary illumination conditions. First, using a bootstrap set consisting of 3D face models, we compute a statistical model for each basis image. During training, given a novel face image under arbitrary illumination, we recover a set of images for this face. We prove that these images are the set of basis images with maximum probability. During testing, we recognize the face for which there exists a weighted combination of basis images that is the closest to the test face image. We provide a series of experiments that achieve high recognition rates, under a wide range of illumination conditions, including multiple sources of illumination. Our method achieves comparable levels of accuracy with methods that have much more onerous training data requirements.


computer vision and pattern recognition | 2012

Two-person interaction detection using body-pose features and multiple instance learning

Kiwon Yun; Jean Honorio; Debaleena Chattopadhyay; Tamara L. Berg; Dimitris Samaras

Human activity recognition has potential to impact a wide range of applications from surveillance to human computer interfaces to content based video retrieval. Recently, the rapid development of inexpensive depth sensors (e.g. Microsoft Kinect) provides adequate accuracy for real-time full-body human tracking for activity recognition applications. In this paper, we create a complex human activity dataset depicting two person interactions, including synchronized video, depth and motion capture data. Moreover, we use our dataset to evaluate various features typically used for indexing and retrieval of motion capture data, in the context of real-time detection of interaction activities via Support Vector Machines (SVMs). Experimentally, we find that the geometric relational features based on distance between all pairs of joints outperforms other feature choices. For whole sequence classification, we also explore techniques related to Multiple Instance Learning (MIL) in which the sequence is represented by a bag of body-pose features. We find that the MIL based classifier outperforms SVMs when the sequences extend temporally around the interaction of interest.


eurographics | 2004

High Resolution Acquisition, Learning and Transfer of Dynamic 3‐D Facial Expressions

Yang Wang; Xiaolei Huang; Chan-Su Lee; Song Zhang; Zhiguo Li; Dimitris Samaras; Dimitris N. Metaxas; Ahmed M. Elgammal; Peisen Huang

Synthesis and re‐targeting of facial expressions is central to facial animation and often involves significant manual work in order to achieve realistic expressions, due to the difficulty of capturing high quality dynamic expression data. In this paper we address fundamental issues regarding the use of high quality dense 3‐D data samples undergoing motions at video speeds, e.g. human facial expressions. In order to utilize such data for motion analysis and re‐targeting, correspondences must be established between data in different frames of the same faces as well as between different faces. We present a data driven approach that consists of four parts: 1) High speed, high accuracy capture of moving faces without the use of markers, 2) Very precise tracking of facial motion using a multi‐resolution deformable mesh, 3) A unified low dimensional mapping of dynamic facial motion that can separate expression style, and 4) Synthesis of novel expressions as a combination of expression styles. The accuracy and resolution of our method allows us to capture and track subtle expression details. The low dimensional representation of motion data in a unified embedding for all the subjects in the database allows for learning the most discriminating characteristics of each individuals expressions as that persons “expression style”. Thus new expressions can be synthesized, either as dynamic morphing between individuals, or as expression transfer from a source face to a target face, as demonstrated in a series of experiments.


computer vision and pattern recognition | 2003

Using multiple cues for hand tracking and model refinement

Shan Lu; Dimitris N. Metaxas; Dimitris Samaras; John Oliensis

We present a model based approach to the integration of multiple cues for tracking high degree of freedom articulated motions and model refinement. We then apply it to the problem of hand tracking using a single camera sequence. Hand tracking is particularly challenging because of occlusions, shading variations, and the high dimensionality of the motion. The novelty of our approach is in the combination of multiple sources of information, which come from edges, optical flow, and shading information in order to refine the model during tracking. We first use a previously formulated generalized version of the gradient-based optical flow constraint, that includes shading flow i.e., the variation of the shading of the object as it rotates with respect to the light source. Using this model we track its complex articulated motion in the presence of shading changes. We use a forward recursive dynamic model to track the motion in response to data derived 3D forces applied to the model. However, due to inaccurate initial shape, the generalized optical flow constraint is violated. We use the error in the generalized optical flow equation to compute generalized forces that correct the model shape at each step. The effectiveness of our approach is demonstrated with experiments on a number of different hand motions with shading changes, rotations and occlusions of significant parts of the hand.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2007

Conformal Geometry and Its Applications on 3D Shape Matching, Recognition, and Stitching

Sen Wang; Yang Wang; Miao Jin; Xianfeng David Gu; Dimitris Samaras

Three-dimensional shape matching is a fundamental issue in computer vision with many applications such as shape registration, 3D object recognition, and classification. However, shape matching with noise, occlusion, and clutter is a challenging problem. In this paper, we analyze a family of quasi-conformal maps including harmonic maps, conformal maps, and least-squares conformal maps with regards to 3D shape matching. As a result, we propose a novel and computationally efficient shape matching framework by using least-squares conformal maps. According to conformal geometry theory, each 3D surface with disk topology can be mapped to a 2D domain through a global optimization and the resulting map is a diffeomorphism, i.e., one-to-one and onto. This allows us to simplify the 3D shape-matching problem to a 2D image-matching problem, by comparing the resulting 2D parametric maps, which are stable, insensitive to resolution changes and robust to occlusion, and noise. Therefore, highly accurate and efficient 3D shape matching algorithms can be achieved by using the above three parametric maps. Finally, the robustness of least-squares conformal maps is evaluated and analyzed comprehensively in 3D shape matching with occlusion, noise, and resolution variation. In order to further demonstrate the performance of our proposed method, we also conduct a series of experiments on two computer vision applications, i.e., 3D face recognition and 3D nonrigid surface alignment and stitching.


computer vision and pattern recognition | 2010

Dense non-rigid surface registration using high-order graph matching

Yun Zeng; Chaohui Wang; Yang Wang; Xianfeng Gu; Dimitris Samaras; Nikos Paragios

In this paper, we propose a high-order graph matching formulation to address non-rigid surface matching. The singleton terms capture the geometric and appearance similarities (e.g., curvature and texture) while the high-order terms model the intrinsic embedding energy. The novelty of this paper includes: 1) casting 3D surface registration into a graph matching problem that combines both geometric and appearance similarities and intrinsic embedding information, 2) the first implementation of high-order graph matching algorithm that solves a non-convex optimization problem, and 3) an efficient two-stage optimization approach to constrain the search space for dense surface registration. Our method is validated through a series of experiments demonstrating its accuracy and efficiency, notably in challenging cases of large and/or non-isometric deformations, or meshes that are partially occluded.


international conference on computer vision | 2007

Real-time Accurate Object Detection using Multiple Resolutions

Wei Zhang; Gregory J. Zelinsky; Dimitris Samaras

We propose a multi-resolution framework inspired by human visual search for general object detection. Different resolutions are represented using a coarse-to-fine feature hierarchy. During detection, the lower resolution features are initially used to reject the majority of negative windows at relatively low cost, leaving a relatively small number of windows to be processed in higher resolutions. This enables the use of computationally more expensive higher resolution features to achieve high detection accuracy. We applied this framework on Histograms of Oriented Gradient (HOG) features for object detection. Our multi-resolution detector produced better performance for pedestrian detection than state-of-the-art methods (Dalal and Triggs, 2005), and was faster during both training and testing. Testing our method on motorbikes and cars from the VOC database revealed similar improvements in both speed and accuracy, suggesting that our approach is suitable for realtime general object detection applications.


Proceedings of the National Academy of Sciences of the United States of America | 2010

Oral methylphenidate normalizes cingulate activity in cocaine addiction during a salient cognitive task

Rita Z. Goldstein; Patricia A. Woicik; Thomas Maloney; Dardo Tomasi; Nelly Alia-Klein; Juntian Shan; Jean Honorio; Dimitris Samaras; Ruiliang Wang; Frank Telang; Gene-Jack Wang; Nora D. Volkow

Anterior cingulate cortex (ACC) hypoactivations during cognitive demand are a hallmark deficit in drug addiction. Methylphenidate (MPH) normalizes cortical function, enhancing task salience and improving associated cognitive abilities, in other frontal lobe pathologies; however, in clinical trials, MPH did not improve treatment outcome in cocaine addiction. We hypothesized that oral MPH will attenuate ACC hypoactivations and improve associated performance during a salient cognitive task in individuals with cocaine-use disorders (CUD). In the current functional MRI study, we used a rewarded drug cue-reactivity task previously shown to be associated with hypoactivations in both major ACC subdivisions (implicated in default brain function) in CUD compared with healthy controls. The task was performed by 13 CUD and 14 matched healthy controls on 2 d: after ingesting a single dose of oral MPH (20 mg) or placebo (lactose) in a counterbalanced fashion. Results show that oral MPH increased responses to this salient cognitive task in both major ACC subdivisions (including the caudal-dorsal ACC and rostroventromedial ACC extending to the medial orbitofrontal cortex) in the CUD. These functional MRI results were associated with reduced errors of commission (a common impulsivity measure) and improved task accuracy, especially during the drug (vs. neutral) cue-reactivity condition in all subjects. The clinical application of such MPH-induced brain-behavior enhancements remains to be tested.


pacific conference on computer graphics and applications | 2002

Estimation of multiple directional light sources for synthesis of mixed reality images

Yang Wang; Dimitris Samaras

We present a new method for the detection and estimation of multiple directional illuminants, using a single image of any object with known geometry and Lambertian reflectance. We use the resulting highly accurate estimates to modify virtually the illumination and geometry of a real scene and produce correctly illuminated mixed reality images. Our method obviates the need to modify the imaged scene by inserting calibration objects of any particular geometry, relying instead on partial knowledge of the geometry of the scene. Thus, the recovered multiple illuminants can be used both for image-based rendering and for shape reconstruction. Our method combines information both from the shading of the object and from shadows cast on the scene by the object. Initially we use a method based on shadows and a method based on shading independently. The shadow based method utilizes brightness variation inside the shadows cast by the object, whereas the shading based method utilizes brightness variation on the directly illuminated portions of the object. We demonstrate how the two sources of information complement each other in a number of occasions. We then describe an approach that integrates the two methods, with results superior to those obtained if the two methods are used separately. The resulting illumination information can be used (i) to render synthetic objects in a real photograph with correct illumination effects, and (ii) to virtually re-light the scene.

Collaboration


Dive into the Dimitris Samaras's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jean Honorio

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Nikos Paragios

Université Paris-Saclay

View shared research outputs
Top Co-Authors

Avatar

Lei Zhang

Stony Brook University

View shared research outputs
Top Co-Authors

Avatar

Rita Z. Goldstein

Icahn School of Medicine at Mount Sinai

View shared research outputs
Top Co-Authors

Avatar

Yun Zeng

Stony Brook University

View shared research outputs
Top Co-Authors

Avatar

Minh Hoai

Stony Brook University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xianfeng Gu

Stony Brook University

View shared research outputs
Researchain Logo
Decentralizing Knowledge