Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Paul Ruvolo is active.

Publication


Featured researches published by Paul Ruvolo.


systems man and cybernetics | 2012

Multilayer Architectures for Facial Action Unit Recognition

Tingfan Wu; Nicholas J. Butko; Paul Ruvolo; Jacob Whitehill; Marian Stewart Bartlett; Javier R. Movellan

In expression recognition and many other computer vision applications, the recognition performance is greatly improved by adding a layer of nonlinear texture filters between the raw input pixels and the classifier. The function of this layer is typically known as feature extraction. Popular filter types for this layer are Gabor energy filters (GEFs) and local binary patterns (LBPs). Recent work [1] suggests that adding a second layer of nonlinear filters on top of the first layer may be beneficial. However, it is unclear what is the best architecture of layers and selection of filters. In this paper, we present a thorough empirical analysis of the performance of single-layer and dual-layer texture-based approaches for action unit recognition. For the single hidden layer case, GEFs perform consistently better than LBPs, which may be due to their robustness to jitter and illumination noise as well as to their ability to encode texture at multiple resolutions. For dual-layer case, we confirm that, while small, the benefit of adding this second layer is reliable and consistent across data sets. Interestingly for this second layer, LBPs appear to perform better than GEFs.


human-robot interaction | 2007

The RUBI project: a progress report

Javier R. Movellan; Fumihide Tanaka; Ian R. Fasel; Cynthia Taylor; Paul Ruvolo; Micah Eckhardt

The goal of the RUBI project is to accelerate progress in the development of social robots by addressing the problem at multiple levels, including the development of a scientific agenda, research methods, formal approaches, software, and hardware. The project is based on the idea that progress will go hand-in-hand with the emergence of a new scientific discipline that focuses on understanding the organization of adaptive behavior in real-time within the environments in which organisms operate. As such, the RUBI project emphasizes the process of design by immersion, i.e., embedding scientists, engineers and robots in everyday life environments so as to have these environments shape the hardware, software, and scientific questions as early as possible in the development process. The focus of the project so far has been on social robots that interact with 18 to 24 month old toddlers as part of their daily activities at the Early Childhood Education Center at the University of California, San Diego. In this document we present an overall assessment of the lessons and progress through year two of the project.


Face and Gesture 2011 | 2011

Action unit recognition transfer across datasets

Tingfan Wu; Nicholas J. Butko; Paul Ruvolo; Jacob Whitehill; Marian Stewart Bartlett; Javier R. Movellan

We explore how CERT [15], a computer expression recognition toolbox trained on a large dataset of spontaneous facial expressions (FFD07), generalizes to a new, previously unseen dataset (FERA). The experiment was unique in that the authors had no access to the test labels, which were guarded as part of the FERA challenge. We show that without any training or special adaptation to the new database, CERT performs better than a baseline method trained exclusively on that database. Best results are achieved by retraining CERT with a combination of old and new data. We also found that the FERA dataset may be too small and idiosyncratic to generalize to other datasets. Training on FERA alone produced good results on FERA but very poor results on FFD07. We reflect on the importance of challenges like this for the future of the field, and discuss suggestions for standardization of future challenges.


Face and Gesture 2011 | 2011

The motion in emotion — A CERT based approach to the FERA emotion challenge

Gwen Littlewort; Jacob Whitehill; Tingfan Wu; Nicholas J. Butko; Paul Ruvolo; Javier R. Movellan; Marian Stewart Bartlett

This paper assesses the performance of measures of facial expression dynamics derived from the Computer Expression Recognition Toolbox (CERT) for classifying emotions in the Facial Expression Recognition and Analysis (FERA) Challenge. The CERT system automatically estimates facial action intensity and head position using learned appearance-based models on single frames of video. CERT outputs were used to derive a representation of the intensity and motion in each video, consisting of the extremes of displacement, velocity and acceleration. Using this representation, emotion detectors were trained on the FERA training examples. Experiments on the released portion of the FERA dataset are presented, as well as results on the blind test. No consideration of subject identity was taken into account in the blind test. The F1 scores were well above the baseline criterion for success.


Neural Networks | 2010

2010 Special Issue: Applying machine learning to infant interaction: The development is in the details

Daniel Messinger; Paul Ruvolo; Naomi V. Ekas; Alan Fogel

The face-to-face interactions of infants and their parents are a model system in which critical communicative abilities emerge. We apply machine learning methods to explore the predictability of infant and mother behavior during interaction with an eye to understanding the preconditions of infant intentionality. Overall, developmental changes were most evident when the probability of specific behaviors was examined in specific interactive contexts. Mothers smiled predictably in response to infant smiles, for example, and infant smile initiations become more predictable over developmental time. Analysis of face-to-face interaction--a tractable model system--promise to pave the way for the construction of virtual and physical agents who are able to interact and develop.


Pattern Recognition Letters | 2010

A learning approach to hierarchical feature selection and aggregation for audio classification

Paul Ruvolo; Ian R. Fasel; Javier R. Movellan

Audio classification typically involves feeding a fixed set of low-level features to a machine learning method, then performing feature aggregation before or after learning. Instead, we jointly learn a selection and hierarchical temporal aggregation of features, achieving significant performance gains.


intelligent robots and systems | 2010

Approaches and databases for online calibration of binaural sound localization for robotic heads

Holger Finger; Shih-Chii Liu; Paul Ruvolo; Javier R. Movellan

In this paper, we evaluate adaptive sound localization algorithms for robotic heads. To this end we built a 3 degree-of-freedom head with two microphones encased in artificial pinnae (outer ears). The geometry of the head and pinnae induce temporal differences in the sound recorded at each microphone. These differences change with the frequency of the sound, location of the sound, and orientation of the robot in a complex manner. To learn the relationship between these auditory differences and the location of a sound source, we applied machine learning methods to a database of different audio source locations and robot head orientations. Our approach achieves a mean error of 2.5 degrees for azimuth and 11 degrees for elevation for estimating the position of an audio source. The impressive results highlight the benefits of a two-stage regression model to make use of the properties of the artificial pinnae for elevation estimation. In this work, the algorithms were trained using ground truth data provided by a motion capture system. We are currently generalizing the approach so that the training signal is provided online based on a real-time face detection and speech detection system.


international conference on robotics and automation | 2008

Auditory mood detection for social and educational robots

Paul Ruvolo; Ian R. Fasel; Javier R. Movellan

Social robots face the fundamental challenge of detecting and adapting their behavior to the current social mood. For example, robots that assist teachers in early education must choose different behaviors depending on whether the children are crying, laughing, sleeping, or singing songs. Interactive robotic applications require perceptual algorithms that both run in real time and are adaptable to the challenging conditions of daily life. This paper explores a novel approach to auditory mood detection which was born out of our experience immersing social robots in classroom environments. We propose a new set of low-level spectral contrast features that extends a class of features which have proven very successful for object recognition in the modern computer vision literature. Features are selected and combined using machine learning approaches so as to make decisions about the ongoing auditory mood. We demonstrate excellent performance on two standard emotional speech databases (the Berlin Emotional Speech [W. Burkhardt et al., 2005], and the ORATOR dataset [H. Quast, 2001]). In addition we establish strong baseline performance for mood detection on a database collected from a social robot immersed in a classroom of 18-24 months old children [J. Movellan er al., 2007]. This approach operates in real time at little computational cost. It has the potential to greatly enhance the effectiveness of social robots in daily life environments.


international conference on development and learning | 2008

Automatic cry detection in early childhood education settings

Paul Ruvolo; Javier R. Movellan

We present results on applying a novel machine learning approach for learning auditory moods in natural environments [1] to the problem of detecting crying episodes in preschool classrooms. The resulting system achieved levels of performance approaching that of human coders and also significantly outperformed previous approaches to this problem [2].


international conference on development and learning | 2008

Building a more effective teaching robot using apprenticeship learning

Paul Ruvolo; Jacob Whitehill; Marjo Virnes; Javier R. Movellan

What defines good teaching? While attributes such as timing, responsiveness to social cues, and pacing of material clearly play a role, it is difficult to create a comprehensive specification of what it means to be a good teacher. On the other hand, it is relatively easy to obtain examples of expert teaching behavior by observing a real teacher. With this inspiration as our guide, we investigated apprenticeship learning methods [1] that use data recorded from expert teachers as a means of improving the teaching abilities of RUBI, a social robot immersed in a classroom of 18-24 month old children. While this approach has achieved considerable success in mechanical control, such as automated helicopter flight [2], until now there has been little work on applying it to the field of social robotics. This paper explores two particular approaches to apprenticeship learning, and analyzes the models of teaching that each approach learns from the data of the human teacher. Empirical results indicate that the apprenticeship learning paradigm, though still nascent in its use in the social robotics field, holds promise, and that our proposed methods can already extract meaningful teaching models from demonstrations of a human expert.

Collaboration


Dive into the Paul Ruvolo's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Eric Eaton

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Tingfan Wu

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Galit Hofree

University of California

View shared research outputs
Top Co-Authors

Avatar

Haitham Bou Ammar

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Matthew E. Taylor

Washington State University

View shared research outputs
Researchain Logo
Decentralizing Knowledge