Michael Johnston
AT&T
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Michael Johnston.
meeting of the association for computational linguistics | 2002
Michael Johnston; SrinivasBangalore; Gunaranjan Vasireddy; Amanda Stent; Patrick Ehlen; Marilyn A. Walker; Steve Whittaker; Preetam Maloor
Mobile interfaces need to allow the user and system to adapt their choice of communication modes according to user preferences, the task at hand, and the physical and social environment. We describe a multimodal application architecture which combines finite-state multimodal language processing, a speech-act based multimodal dialogue manager, dynamic multimodal output generation, and user-tailored text planning to enable rapid prototyping of multimodal interfaces with flexible input and adaptive output. Our testbed application MATCH (Multimodal Access To City Help) provides a mobile multimodal speech-pen interface to restaurant and sub-way information for New York City.
Computational Linguistics | 2009
Srinivas Bangalore; Michael Johnston
Multimodal grammars provide an effective mechanism for quickly creating integration and understanding capabilities for interactive systems supporting simultaneous use of multiple input modalities. However, like other approaches based on hand-crafted grammars, multimodal grammars can be brittle with respect to unexpected, erroneous, or disfluent input. In this article, we show how the finite-state approach to multimodal language processing can be extended to support multimodal applications combining speech with complex freehand pen input, and evaluate the approach in the context of a multimodal conversational system (MATCH). We explore a range of different techniques for improving the robustness of multimodal integration and understanding. These include techniques for building effective language models for speech recognition when little or no multimodal training data is available, and techniques for robust multimodal understanding that draw on classification, machine translation, and sequence edit methods. We also explore the use of edit-based methods to overcome mismatches between the gesture stream and the speech stream.
meeting of the association for computational linguistics | 2004
Michael Johnston; Srinivas Bangalore
Multimodal interfaces provide more flexible and compelling interaction and can enable public information kiosks to support more complex tasks for a broader community of users. MATCHKiosk is a multimodal interactive city guide which provides users with the freedom to interact using speech, pen, touch or multimodal inputs. The system responds by generating multimodal presentations that synchronize synthetic speech with a life-like virtual agent and dynamically generated graphics.
annual meeting of the special interest group on discourse and dialogue | 2014
Michael Johnston; John Chen; Patrick Ehlen; Hyuckchul Jung; Jay Lieske; Aarthi M. Reddy; Ethan O. Selfridge; Svetlana Stoyanchev; Brant J. Vasilieff; Jay Gordon Wilpon
The Multimodal Virtual Assistant (MVA) is an application that enables users to plan an outing through an interactive multimodal dialog with a mobile device. MVA demonstrates how a cloud-based multimodal language processing infrastructure can support mobile multimodal interaction. This demonstration will highlight incremental recognition, multimodal speech and gesture input, contextually-aware language understanding, and the targeted clarification of potentially incorrect segments within user input.
IEEE Signal Processing Magazine | 2011
Junlan Feng; Michael Johnston; Srinivas Bangalore
With the widespread adoption of high-speed wireless networks symbiotically complemented by the burgeoning demand for smart mobile devices, access to the Internet is evolving from personal computers (PCs) to mobile devices. In this article, we highlight the characteristics of mobile search, discuss the state of speech-based mobile search, and present opportunities for exploiting multimodal interaction to optimize the efficiency of mobile search.
spoken language technology workshop | 2010
Marcelo Worsley; Michael Johnston
Through the growing popularity of voice-enabled search, multimodal applications are finally starting to get into the hands of consumers. However, these applications are principally for mobile platforms and generally involve highly-moded interaction where the user has to click or hold a button in order to speak. Significant technical challenges remain in bringing multimodal interaction to other environments such as smart living rooms and classrooms, where users speech and gesture is directed toward large displays or interactive kiosks and the microphone and other sensors are ‘always on’. In this demonstration, we present a framework combining low cost hardware and open source software that lowers the barrier of entry for exploration of multimodal interaction in smart environments. Specifically, we will demonstrate the combination of infrared tracking, face detection, and open microphone speech recognition for media search (magicTV) and map navigation (magicMap).
interaction design and children | 2011
Marcelo Worsley; Michael Johnston; Paulo Blikstein
In this paper, we present an application framework for enabling education practitioners and researchers to develop interactive, multi-modal applications. These applications can be designed using typical HTML programming, and will enable a larger audience to make applications that incorporate speech recognition, gesture recognition and engagement detection. The application framework uses open-source software and inexpensive hardware that supports both multi-touch and multi-user capabilities.
international conference on multimodal interfaces | 2015
Ethan O. Selfridge; Michael Johnston
Interact is a mobile virtual assistant that uses multimodal dialog to enable an interactive concierge experience over multiple application domains including hotel, restaurants, events, and TV search. Interact demonstrates how multi- modal interaction combined with conversational dialog en- ables a richer and more natural user experience. This demonstration will highlight incremental recognition and under- standing, multimodal speech and gesture input, context track- ing over multiple simultaneous domains, and the use of multimodal interface techniques to enable disambiguation of erors and online personalization.
Computational Linguistics | 2003
Michael Johnston
We are a species of storytellers. We tell stories in many different ways. In ancient legends from Gilgamesh to the Odyssey to the Pentateuch, we set down out pre-literate lore. In proverbs, aphorisms and epigrams, we condense our wisdom. We write poems, romances and detective novels. We write historical romances and “how-to” books that explain how to find love, quit booze, get a better job or fix up an old car. When threatened, we growl ominously, whimper pathetically or compose legal defence briefs. Sometimes, especially when we are ambitiously seeking the truth, we try science.
Computational Linguistics | 2003
Michael Johnston
We are a species of storytellers. We tell stories in many different ways. In ancient legends from Gilgamesh to the Odyssey to the Pentateuch, we set down out pre-literate lore. In proverbs, aphorisms and epigrams, we condense our wisdom. We write poems, romances and detective novels. We write historical romances and “how-to” books that explain how to find love, quit booze, get a better job or fix up an old car. When threatened, we growl ominously, whimper pathetically or compose legal defence briefs. Sometimes, especially when we are ambitiously seeking the truth, we try science.