Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jesse Thomason is active.

Publication


Featured researches published by Jesse Thomason.


The International Journal of Robotics Research | 2017

BWIBots: A platform for bridging the gap between AI and human–robot interaction research

Piyush Khandelwal; Shiqi Zhang; Jivko Sinapov; Matteo Leonetti; Jesse Thomason; Fangkai Yang; Ilaria Gori; Maxwell Svetlik; Priyanka Khante; Vladimir Lifschitz; Jake K. Aggarwal; Raymond J. Mooney; Peter Stone

Recent progress in both AI and robotics have enabled the development of general purpose robot platforms that are capable of executing a wide variety of complex, temporally extended service tasks in open environments. This article introduces a novel, custom-designed multi-robot platform for research on AI, robotics, and especially human–robot interaction for service robots. Called BWIBots, the robots were designed as a part of the Building-Wide Intelligence (BWI) project at the University of Texas at Austin. The article begins with a description of, and justification for, the hardware and software design decisions underlying the BWIBots, with the aim of informing the design of such platforms in the future. It then proceeds to present an overview of various research contributions that have enabled the BWIBots to better (a) execute action sequences to complete user requests, (b) efficiently ask questions to resolve user requests, (c) understand human commands given in natural language, and (d) understand human intention from afar. The article concludes with a look forward towards future research opportunities and applications enabled by the BWIBot platform.


meeting of the association for computational linguistics | 2017

Guiding Interaction Behaviors for Multi-modal Grounded Language Learning

Jesse Thomason; Jivko Sinapov; Raymond J. Mooney

Multi-modal grounded language learning connects language predicates to physical properties of objects in the world. Sensing with multiple modalities, such as audio, haptics, and visual colors and shapes while performing interaction behaviors like lifting, dropping, and looking on objects enables a robot to ground non-visual predicates like “empty” as well as visual predicates like “red”. Previous work has established that grounding in multi-modal space improves performance on object retrieval from human descriptions. In this work, we gather behavior annotations from humans and demonstrate that these improve language grounding performance by allowing a system to focus on relevant behaviors for words like “white” or “half-full” that can be understood by looking or lifting, respectively. We also explore adding modality annotations (whether to focus on audio or haptics when performing a behavior), which improves performance, and sharing information between linguistically related predicates (if “green” is a color, “white” is a color), which improves grounding recall but at the cost of precision.


international joint conference on artificial intelligence | 2018

Multi-modal Predicate Identification using Dynamically Learned Robot Controllers

Saeid Amiri; Suhua Wei; Shiqi Zhang; Jivko Sinapov; Jesse Thomason; Peter Stone

Intelligent robots frequently need to explore the objects in their working environments. Modern sensors have enabled robots to learn object properties via perception of multiple modalities. However, object exploration in the real world poses a challenging trade-off between information gains and exploration action costs. Mixed observability Markov decision process (MOMDP) is a framework for planning under uncertainty, while accounting for both fully and partially observable components of the state. Robot perception frequently has to face such mixed observability. This work enables a robot equipped with an arm to dynamically construct query-oriented MOMDPs for multi-modal predicate identification (MPI) of objects. The robot’s behavioral policy is learned from two datasets collected using real robots. Our approach enables a robot to explore object properties in a way that is significantly faster while improving accuracies in comparison to existing methods that rely on handcoded exploration strategies.


Archive | 2018

Continually improving grounded natural language understanding through human-robot dialog

Jesse Thomason

Natural language understanding for robotics can require substantial domainand platform-specific engineering. For example, for mobile robots to pick-and-place objects in an environment to satisfy human commands, we can specify the language humans use to issue such commands, and connect concept words like red can to physical object properties. One way to alleviate this engineering for a new domain is to enable robots in human environments to adapt dynamically— continually learning new language constructions and perceptual concepts. In this work, we present an end-to-end pipeline for translating natural language commands to discrete robot actions, and use clarification dialogs to jointly improve language parsing and concept grounding. We train and evaluate this agent in a virtual setting on Amazon Mechanical Turk, and we transfer the learned agent to a physical robot platform to demonstrate it in the real world.


international joint conference on artificial intelligence | 2017

Multi-Modal Word Synset Induction

Jesse Thomason; Raymond J. Mooney

A word in natural language can be polysemous, having multiple meanings, as well as synonymous, meaning the same thing as other words. Word sense induction attempts to find the senses of polysemous words. Synonymy detection attempts to find when two words are interchangeable. We combine these tasks, first inducing word senses and then detecting similar senses to form word-sense synonym sets (synsets) in an unsupervised fashion. Given pairs of images and text with noun phrase labels, we perform synset induction to produce collections of underlying concepts described by one or more noun phrases. We find that considering multimodal features from both visual and textual context yields better induced synsets than using either context alone. Human evaluations show that our unsupervised, multi-modally induced synsets are comparable in quality to annotation-assisted ImageNet synsets, achieving about 84% of ImageNet synsets’ approval.


international conference on computational linguistics | 2014

Integrating Language and Vision to Generate Natural Language Descriptions of Videos in the Wild

Jesse Thomason; Subhashini Venugopalan; Sergio Guadarrama; Kate Saenko; Raymond J. Mooney


international conference on artificial intelligence | 2015

Learning to interpret natural language commands through human-robot dialog

Jesse Thomason; Shiqi Zhang; Raymond J. Mooney; Peter Stone


international joint conference on artificial intelligence | 2016

Learning multi-modal grounded linguistic semantics by playing I Spy

Jesse Thomason; Jivko Sinapov; Maxwell Svetlik; Peter Stone; Raymond J. Mooney


conference of the european chapter of the association for computational linguistics | 2017

Integrated Learning of Dialog Strategies and Semantic Parsing

Aishwarya Padmakumar; Jesse Thomason; Raymond J. Mooney


Conference on Robot Learning | 2017

Opportunistic Active Learning for Grounding Natural Language Descriptions.

Jesse Thomason; Aishwarya Padmakumar; Jivko Sinapov; Justin Hart; Peter Stone; Raymond J. Mooney

Collaboration


Dive into the Jesse Thomason's collaboration.

Top Co-Authors

Avatar

Raymond J. Mooney

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Peter Stone

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Shiqi Zhang

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Maxwell Svetlik

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Fangkai Yang

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Ilaria Gori

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

Jake K. Aggarwal

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar

James G. Scott

University of Texas at Austin

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge