Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Elizabeth S. Redden is active.

Publication


Featured researches published by Elizabeth S. Redden.


international conference on multimodal interfaces | 2006

Comparing the effects of visual-auditory and visual-tactile feedback on user performance: a meta-analysis

Jennifer L. Burke; Matthew S. Prewett; Ashley A. Gray; Liuquin Yang; Frederick R. B. Stilson; Michael D. Coovert; Linda R. Elliot; Elizabeth S. Redden

In a meta-analysis of 43 studies, we examined the effects of multimodal feedback on user performance, comparing visual-auditory and visual-tactile feedback to visual feedback alone. Results indicate that adding an additional modality to visual feedback improves performance overall. Both visual-auditory feedback and visual-tactile feedback provided advantages in reducing reaction times and improving performance scores, but were not effective in reducing error rates. Effects are moderated by task type, workload, and number of tasks. Visual-auditory feedback is most effective when a single task is being performed (g = .87), and under normal workload conditions (g = .71). Visual-tactile feedback is more effective when multiple tasks are begin performed (g = .77) and workload conditions are high (g = .84). Both types of multimodal feedback are effective for target acquisition tasks; but vary in effectiveness for other task types. Implications for practice and research are discussed.


international conference on multimodal interfaces | 2006

The benefits of multimodal information: a meta-analysis comparing visual and visual-tactile feedback

Matthew S. Prewett; Liuquin Yang; Frederick R. B. Stilson; Ashley A. Gray; Michael D. Coovert; Jennifer L. Burke; Elizabeth S. Redden; Linda R. Elliot

Information display systems have become increasingly complex and more difficult for human cognition to process effectively. Based upon Wickens Multiple Resource Theory (MRT), information delivered using multiple modalities (i.e., visual and tactile) could be more effective than communicating the same information through a single modality. The purpose of this meta-analysis is to compare user effectiveness when using visual-tactile task feedback (a multimodality) to using only visual task feedback (a single modality). Results indicate that using visual-tactile feedback enhances task effectiveness more so than visual feedback (g = .38). When assessing different criteria, visual-tactile feedback is particularly effective at reducing reaction time (g = .631) and increasing performance (g = .618). Follow up moderator analyses indicate that visual-tactile feedback is more effective when workload is high (g = .844) and multiple tasks are being performed (g = .767). Implications of results are discussed in the paper.


Ergonomics in Design | 2015

Tactile Cuing to Augment Multisensory Human–Machine Interaction.

Peter A. Hancock; Ben D. Lawson; Roger W. Cholewiak; Linda R. Elliott; Johannes Bernardus Fransiscus van Erp; Bruce J. P. Mortimer; Angus H. Rupert; Elizabeth S. Redden

Tactile displays promise to improve the information-processing capacity of operators, especially when used in conjunction with visual and auditory displays. In this article, we describe current applications and future directions in tactile cuing.


systems man and cybernetics | 2008

A Tool for the Accumulation and Evaluation of Multimodal Research

Michael D. Coovert; Ashley A. G. Walvoord; Linda R. Elliott; Elizabeth S. Redden

A surge of interest exists in multimodal research and interfaces. This is due, at least in part, to an exponential increase in the amount and type of information that can be presented to a user. When a great deal of information is presented via a single sensory modality, it can exceed the operators capacity to manage the information efficiently, generating cognitive overload. As a consequence, the users performance becomes susceptible to slower response times, loss of situational awareness, faulty decision making, and execution errors. Researchers and designers have responded to these issues with the development and application of multimodal information displays. The cross-disciplinary flavor of multimodal applications presents a challenge to the accumulation, evaluation, and dissemination of relevant research. We describe the development of a taxonomy for the evaluation and comparison of multimodal display research studies, and the implementation of the taxonomy into a database: the Multimodal Query System (MQueS).


Archive | 2007

Remote Tactile Displays for Future Soldiers

Richard D. Gilson; Elizabeth S. Redden; Linda R. Elliott


Archive | 2006

Comparison of Army Hand and Arm Signals to a Covert Tactile Communication System in a Dynamic Environment

Rodger A. Pettitt; Elizabeth S. Redden; Christian B. Carstens


Archive | 2007

Multimodal Guidance for Land Navigation

Linda R. Elliott; Maaike Duistermaat; Elizabeth S. Redden; Jan B. F. van Erp


Archive | 2006

Tactile Guidance for Land Navigation

Linda R. Elliot; Elizabeth S. Redden; Rodger A. Pettitt; Christian B. Carstens; Jan B. F. van Erp; Maaike Duistermaat


Archive | 2005

Effects of Visual, Auditory, and Tactile Cues on Army Platoon Leader Decision Making

Andrea S. Krausman; Linda R. Elliott; Elizabeth S. Redden; Plamen V. Petrov


Archive | 2008

Scalability of Robotic Controllers: An Evaluation of Controller Options

Rodger A. Pettitt; Elizabeth S. Redden; Christian B. Carstens

Collaboration


Dive into the Elizabeth S. Redden's collaboration.

Top Co-Authors

Avatar

Michael D. Coovert

University of South Florida

View shared research outputs
Top Co-Authors

Avatar

Ashley A. Gray

University of South Florida

View shared research outputs
Top Co-Authors

Avatar

Linda R. Elliot

Science Applications International Corporation

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jennifer L. Burke

University of South Florida

View shared research outputs
Top Co-Authors

Avatar

Liuquin Yang

University of South Florida

View shared research outputs
Top Co-Authors

Avatar

Matthew S. Prewett

University of South Florida

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Erin M. Jackson

University of South Florida

View shared research outputs
Top Co-Authors

Avatar

Peter A. Hancock

University of Central Florida

View shared research outputs
Researchain Logo
Decentralizing Knowledge