Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Matthew W. Hoffman is active.

Publication


Featured researches published by Matthew W. Hoffman.


Neural Networks | 2006

2006 Special issue: A probabilistic model of gaze imitation and shared attention

Matthew W. Hoffman; David B. Grimes; Aaron P. Shon; Rajesh P. N. Rao

An important component of language acquisition and cognitive learning is gaze imitation. Infants as young as one year of age can follow the gaze of an adult to determine the object the adult is focusing on. The ability to follow gaze is a precursor to shared attention, wherein two or more agents simultaneously focus their attention on a single object in the environment. Shared attention is a necessary skill for many complex, natural forms of learning, including learning based on imitation. This paper presents a probabilistic model of gaze imitation and shared attention that is inspired by Meltzoff and Moores AIM model for imitation in infants. Our model combines a probabilistic algorithm for estimating gaze vectors with bottom-up saliency maps of visual scenes to produce maximum a posteriori (MAP) estimates of objects being looked at by an observed instructor. We test our model using a robotic system involving a pan-tilt camera head and show that combining saliency maps with gaze estimates leads to greater accuracy than using gaze alone. We additionally show that the system can learn instructor-specific probability distributions over objects, leading to increasing gaze accuracy over successive interactions with the instructor. Our results provide further support for probabilistic models of imitation and suggest new ways of implementing robotic systems that can interact with humans over an extended period of time.


Langmuir | 2012

Nanoscale clustering of carbohydrate thiols in mixed self-assembled monolayers on gold.

Faifan Tantakitti; Jesse Burk-Rafel; Fang Cheng; Robert Egnatchik; Tate Owen; Matthew W. Hoffman; Dirk N. Weiss; Daniel M. Ratner

Self-assembled monolayers (SAMs) bearing pendant carbohydrate functionality are frequently employed to tailor glycan-specific bioactivity onto gold substrates. The resulting glycoSAMs are valuable for interrogating glycan-mediated biological interactions via surface analytical techniques, microarrays, and label-free biosensors. GlycoSAM composition can be readily modified during assembly by using mixed solutions containing thiolated species, including carbohydrates, oligo(ethylene glycol) (OEG), and other inert moieties. This intrinsic tunability of the self-assembled system is frequently used to optimize bioavailability and antibiofouling properties of the resulting SAM. However, until now, our nanoscale understanding of the behavior of these mixed glycoSAMs has lacked detail. In this study, we examined the time-dependent clustering of mixed sugar + OEG glycoSAMs on ultraflat gold substrates. Composition and surface morphologic changes in the monolayers were analyzed by X-ray photoelectron spectroscopy (XPS) and atomic force microscopy (AFM), respectively. We provide evidence that the observed clustering is consistent with a phase separation process in which surface-bound glycans self-associate to form dense glycoclusters within the monolayer. These observations have significant implications for the construction of mixed glycoSAMs for use in biosensing and glycomics applications.


european workshop on reinforcement learning | 2011

Regularized least squares temporal difference learning with nested ℓ 2 and ℓ 1 penalization

Matthew W. Hoffman; Alessandro Lazaric; Mohammad Ghavamzadeh; Rémi Munos

The construction of a suitable set of features to approximate value functions is a central problem in reinforcement learning (RL). A popular approach to this problem is to use high-dimensional feature spaces together with least-squares temporal difference learning (LSTD). Although this combination allows for very accurate approximations, it often exhibits poor prediction performance because of overfitting when the number of samples is small compared to the number of features in the approximation space. In the linear regression setting, regularization is commonly used to overcome this problem. In this paper, we review some regularized approaches to policy evaluation and we introduce a novel scheme (L 21 ) which uses l2 regularization in the projection operator and an l1 penalty in the fixed-point step. We show that such formulation reduces to a standard Lasso problem. As a result, any off-the-shelf solver can be used to compute its solution and standardization techniques can be applied to the data. We report experimental results showing that L 21 is effective in avoiding overfitting and that it compares favorably to existing l1 regularized methods.


international conference on robotics and automation | 2005

Probabilistic Gaze Imitation and Saliency Learning in a Robotic Head

Aaron P. Shon; David B. Grimes; Chris L. Baker; Matthew W. Hoffman; Shengli Zhou; Rajesh P. N. Rao

Imitation is a powerful mechanism for transferring knowledge from an instructor to a naïve observer, one that is deeply contingent on a state of shared attention between these two agents. In this paper we present Bayesian algorithms that implement the core of an imitation learning framework. We use gaze imitation, coupled with task-dependent saliency learning, to build a state of shared attention between the instructor and observer. We demonstrate the performance of our algorithms in a gaze following and saliency learning task implemented on an active vision robotic head. Our results suggest that the ability to follow gaze and learn instructor-and task-specific saliency models could play a crucial role in building systems capable of complex forms of human-robot interaction.


Journal of Machine Learning Research | 2016

A general framework for constrained Bayesian optimization using information-based search

José Miguel Hernández-Lobato; Michael A. Gelbart; Ryan P. Adams; Matthew W. Hoffman; Zoubin Ghahramani

We present an information-theoretic framework for solving global black-box optimization problems that also have black-box constraints. Of particular interest to us is to efficiently solve problems with decoupled constraints, in which subsets of the objective and constraint functions may be evaluated independently. For example, when the objective is evaluated on a CPU and the constraints are evaluated independently on a GPU. These problems require an acquisition function that can be separated into the contributions of the individual function evaluations. We develop one such acquisition function and call it Predictive Entropy Search with Constraints (PESC). PESC is an approximation to the expected information gain criterion and it compares favorably to alternative approaches based on improvement in several synthetic and real-world problems. In addition to this, we consider problems with a mix of functions that are fast and slow to evaluate. These problems require balancing the amount of time spent in the meta-computation of PESC and in the actual evaluation of the target objective. We take a bounded rationality approach and develop a partial update for PESC which trades o_ accuracy against speed. We then propose a method for adaptively switching between the partial and full updates for PESC. This allows us to interpolate between versions of PESC that are efficient in terms of function evaluations and those that are efficient in terms of wall-clock time. Overall, we demonstrate that PESC is an effective algorithm that provides a promising direction towards a unified solution for constrained Bayesian optimization.


iberian conference on pattern recognition and image analysis | 2009

Inference and Learning for Active Sensing, Experimental Design and Control

Hendrik Kueck; Matthew W. Hoffman; Arnaud Doucet; Nando de Freitas

In this paper we argue that maximum expected utility is a suitable framework for modeling a broad range of decision problems arising in pattern recognition and related fields. Examples include, among others, gaze planning and other active vision problems, active learning, sensor and actuator placement and coordination, intelligent human-computer interfaces, and optimal control. Following this remark, we present a common inference and learning framework for attacking these problems. We demonstrate this approach on three examples: (i) active sensing with nonlinear, non-Gaussian, continuous models, (ii) optimal experimental design to discriminate among competing scientific models, and (iii) nonlinear optimal control.


Archive | 2013

Decision making with inference and learning methods

Matthew W. Hoffman

In this work we consider probabilistic approaches to sequential decision making. The ultimate goal is to provide methods by which decision making problems can be attacked by approaches and algorithms originally built for probabilistic inference. This in turn allows us to directly apply a wide variety of popular, practical algorithms to these tasks. In Chapter 1 we provide an overview of the general problem of sequential decision making and a broad description of various solution methods. Much of the remaining work of this thesis then proceeds by relying upon probabilistic reinterpretations of the decision making process. This strategy of reducing learning problems to simpler inference tasks has been shown to be very fruitful in much of machine learning, and we expect similar improvements to arise in the control and reinforcement learning fields. The approaches of Chapters 2–3 build upon the framework of [Toussaint and Storkey, 2006] in reformulating the solution of Markov decision processes instead as maximum-likelihood estimation in an equivalent probabilistic model. In Chapter 2 we utilize this framework to construct an Expectation Maximization algorithm for continuous, linear-Gaussian models with mixture-of-Gaussian rewards. This approach extends popular linearquadratic reward models to a much more general setting. We also show how to extend this probabilistic framework to continuous time processes. Chapter 3 further builds upon these methods to introduce a Bayesian approach to policy search using Markov chain Monte Carlo. In Chapter 4 we depart from the setting of direct policy search and instead consider value function estimation. In particular we utilize leastsquares temporal difference learn-


neural information processing systems | 2016

Learning to learn by gradient descent by gradient descent

Marcin Andrychowicz; Misha Denil; Sergio Gómez; Matthew W. Hoffman; David Pfau; Tom Schaul; Brendan Shillingford; Nando de Freitas


neural information processing systems | 2014

Predictive Entropy Search for Efficient Global Optimization of Black-box Functions

José Miguel Hernández-Lobato; Matthew W. Hoffman; Zoubin Ghahramani


international conference on machine learning | 2011

Finite-Sample Analysis of Lasso-TD

Mohammad Ghavamzadeh; Alessandro Lazaric; Matthew W. Hoffman; R mi Munos

Collaboration


Dive into the Matthew W. Hoffman's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Bobak Shahriari

University of British Columbia

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge