Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Boris Schauerte is active.

Publication


Featured researches published by Boris Schauerte.


european conference on computer vision | 2012

Quaternion-Based spectral saliency detection for eye fixation prediction

Boris Schauerte; Rainer Stiefelhagen

In recent years, several authors have reported that spectral saliency detection methods provide state-of-the-art performance in predicting human gaze in images (see, e.g., [1---3]). We systematically integrate and evaluate quaternion DCT- and FFT-based spectral saliency detection [3,4], weighted quaternion color space components [5], and the use of multiple resolutions [1]. Furthermore, we propose the use of the eigenaxes and eigenangles for spectral saliency models that are based on the quaternion Fourier transform. We demonstrate the outstanding performance on the Bruce-Tsotsos (Toronto), Judd (MIT), and Kootstra- Schomacker eye-tracking data sets.


workshop on applications of computer vision | 2012

Predicting human gaze using quaternion DCT image signature saliency and face detection

Boris Schauerte; Rainer Stiefelhagen

We combine and extend the previous work on DCT-based image signatures and face detection to determine the visual saliency. To this end, we transfer the scalar definition of image signatures to quaternion images and thus introduce a novel saliency method using quaternion type-II DCT image signatures. Furthermore, we use MCT-based face detection to model the important influence of faces on the visual saliency using rotated elliptical Gaussian weight functions and evaluate several integration schemes. In order to demonstrate the performance of the proposed methods, we evaluate our approach on the Bruce-Tsotsos (Toronto) [2] and Cerf (FIFA) [3] benchmark eye-tracking data sets. Additionally, we present evaluation results on the Bruce-Tsotsos data set of the most important spectral saliency approaches. We achieve state-of-the-art results in terms of the well-established area under curve (AUC) measure on the Bruce-Tsotsos data set and come close to the ideal AUC on the Cerf data set - with less than one millisecond to calculate the bottom-up QDCT saliency map.


international conference on multimodal interfaces | 2010

Focusing computational visual attention in multi-modal human-robot interaction

Boris Schauerte; Gernot A. Fink

Identifying verbally and non-verbally referred-to objects is an important aspect of human-robot interaction. Most importantly, it is essential to achieve a joint focus of attention and, thus, a natural interaction behavior. In this contribution, we introduce a saliency-based model that reflects how multi-modal referring acts influence the visual search, i.e. the task to find a specific object in a scene. Therefore, we combine positional information obtained from pointing gestures with contextual knowledge about the visual appearance of the referred-to object obtained from language. The available information is then integrated into a biologically-motivated saliency model that forms the basis for visual search. We prove the feasibility of the proposed approach by presenting the results of an experimental evaluation.


international conference on pattern recognition | 2014

Manifold Alignment for Person Independent Appearance-Based Gaze Estimation

Timo Schneider; Boris Schauerte; Rainer Stiefelhagen

We show that dually supervised manifold embedding can improve the performance of machine learning based person-independent and thus calibration-free gaze estimation. For this purpose, we perform a manifold embedding for each person in the training dataset and then learn a linear transformation that aligns the individual, person-dependent manifolds. We evaluate the effect of manifold alignment on the recently presented Columbia dataset, where we analyze the influence on 6 regression methods and 8 feature variants. Using manifold alignment, we are able to improve the person-independent gaze estimation performance by up to 31.2 % compared to the best approach without manifold alignment.


intelligent robots and systems | 2011

Multimodal saliency-based attention for object-based scene analysis

Boris Schauerte; Benjamin Kühn; Kristian Kroschel; Rainer Stiefelhagen

Multimodal attention is a key requirement for humanoid robots in order to navigate in complex environments and act as social, cognitive human partners. To this end, robots have to incorporate attention mechanisms that focus the processing on the potentially most relevant stimuli while controlling the sensor orientation to improve the perception of these stimuli. In this paper, we present our implementation of audio-visual saliency-based attention that we integrated in a system for knowledge-driven audio-visual scene analysis and object-based world modeling. For this purpose, we introduce a novel isophote-based method for proto-object segmentation of saliency maps, a surprise-based auditory saliency definition, and a parametric 3-D model for multimodal saliency fusion. The applicability of the proposed system is demonstrated in a series of experiments.


international conference on multimodal interfaces | 2009

Multi-modal and multi-camera attention in smart environments

Boris Schauerte; Jan Richarz; Thomas Plötz; Christian Thurau; Gernot A. Fink

This paper considers the problem of multi-modal saliency and attention. Saliency is a cue that is often used for directing attention of a computer vision system, e.g., in smart environments or for robots. Unlike the majority of recent publications on visual/audio saliency, we aim at a well grounded integration of several modalities. The proposed framework is based on fuzzy aggregations and offers a flexible, plausible, and efficient way for combining multi-modal saliency information. Besides incorporating different modalities, we extend classical 2D saliency maps to multi-camera and multi-modal 3D saliency spaces. For experimental validation we realized the proposed system within a smart environment. The evaluation took place for a demanding setup under real-life conditions, including focus of attention selection for multiple subjects and concurrently active modalities.


international conference on computers helping people with special needs | 2012

An assistive vision system for the blind that helps find lost things

Boris Schauerte; Manuel Martinez; Angela Constantinescu; Rainer Stiefelhagen

We present a computer vision system that helps blind people find lost objects. To this end, we combine color- and SIFT-based object detection with sonification to guide the hand of the user towards potential target object locations. This way, we are able to guide the users attention and effectively reduce the space in the environment that needs to be explored. We verified the suitability of the proposed system in a user study.


intelligent robots and systems | 2010

Saliency-based identification and recognition of pointed-at objects

Boris Schauerte; Jan Richarz; Gernot A. Fink

When persons interact, non-verbal cues are used to direct the attention of persons towards objects of interest. Achieving joint attention this way is an important aspect of natural communication. Most importantly, it allows to couple verbal descriptions with the visual appearance of objects, if the referred-to object is non-verbally indicated. In this contribution, we present a system that utilizes bottom-up saliency and pointing gestures to efficiently identify pointed-at objects. Furthermore, the system focuses the visual attention by steering a pan-tilt-zoom camera towards the object of interest and thus provides a suitable model-view for SIFT-based recognition and learning. We demonstrate the practical applicability of the proposed system through experimental evaluation in different environments with multiple pointers and objects.


digital image computing: techniques and applications | 2010

Web-Based Learning of Naturalized Color Models for Human-Machine Interaction

Boris Schauerte; Gernot A. Fink

In recent years, natural verbal and non-verbal human-robot interaction has attracted an increasing interest. Therefore, models for robustly detecting and describing visual attributes of objects such as, e.g., colors are of great importance. However, in order to learn robust models of visual attributes, large data sets are required. Based on the idea to overcome the shortage of annotated training data by acquiring images from the Internet, we propose a method for robustly learning natural color models. Its novel aspects with respect to prior art are: firstly, a randomized HSL transformation that reflects the slight variations and noise of colors observed in real-world imaging sensors, secondly, a probabilistic ranking and selection of the training samples, which removes a considerable amount of outliers from the training data. These two techniques allow us to estimate robust color models that better resemble the variances seen in real world images. The advantages of the proposed method over the current state-of-the-art technique using the training data without proper transformation and selection are confirmed in experimental evaluations. In combination, for models learned with pLSA-bg and HSL, the proposed techniques reduce the amount of mislabeled objects by 19.87% on the well-known E-Bay data set.


intelligent robots and systems | 2014

“Look at this!” learning to guide visual saliency in human-robot interaction

Boris Schauerte; Rainer Stiefelhagen

We learn to direct visual saliency in multimodal (i.e., pointing gestures and spoken references) human-robot interaction to highlight and segment arbitrary referent objects. For this purpose, we train a conditional random field to integrate features that reflect low-level visual saliency, the likelihood of salient objects, the probability that a given pixel is pointed at, and - if available - spoken information about the target objects visual appearance. As such, this work integrates several of our ideas and approaches, ranging from multi-scale spectral saliency detection, spatially debiased salient object detection, computational attention in human-robot interaction to learning robust color term models. We demonstrate that this machine learning driven integration outperforms the previously reported results on two datasets, one dataset without and one with spoken object references. In summary, for automatically detected pointing gestures and automatically extracted object references, our approach improves the rate at which the correct object is included in the initial focus of attention by 10.37% in the absence and 25.21% in the presence of spoken target object information.

Collaboration


Dive into the Boris Schauerte's collaboration.

Top Co-Authors

Avatar

Rainer Stiefelhagen

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Gernot A. Fink

Technical University of Dortmund

View shared research outputs
Top Co-Authors

Avatar

Manuel Martinez

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Torsten Wörtwein

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Daniel Koester

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Angela Constantinescu

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Benjamin Kühn

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jan Richarz

Technical University of Dortmund

View shared research outputs
Top Co-Authors

Avatar

Karin Müller

Karlsruhe Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Thomas Plötz

Georgia Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge