Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ha Trinh is active.

Publication


Featured researches published by Ha Trinh.


conference on computers and accessibility | 2014

Phoneme-based predictive text entry interface

Ha Trinh; Annalu Waller; Keith Vertanen; Per Ola Kristensson; Vicki L. Hanson

Phoneme-based text entry provides an alternative typing method for nonspeaking individuals who often experience difficulties in orthographic spelling. In this paper, we investigate the application of rate enhancement strategies to improve the user performance of phoneme-based text entry systems. We have developed a phoneme-based predictive typing system, which employs statistical language modeling techniques to dynamically reduce the phoneme search space and offer accurate word predictions. Results of a case study with a nonspeaking participant demonstrated that our rate enhancement strategies led to improved text entry speed and error rates.


human factors in computing systems | 2015

DynamicDuo: Co-presenting with Virtual Agents

Ha Trinh; Lazlo Ring; Timothy W. Bickmore

The quality of most professional oral presentations is often poor, owing to a number of factors, including public speaking anxiety. We present DynamicDuo, a system that uses an automated, life-sized, animated agent to help inexperienced speakers deliver their presentations in front of an audience. The design of the system was informed by an analysis of TED talks given by two human presenters to identify the most common dual-presentation formats and transition behaviors used. In a within-subjects study (N=12) comparing co-presenting with DynamicDuo against solo-presenting with conventional presentation software, we demonstrated that our system led to significant improvements in public speaking anxiety and speaking confidence for non-native English speakers. Judges who viewed videotapes of these presentations rated those with DynamicDuo significantly higher on speech quality and overall presentation quality for all presenters.


conference on computers and accessibility | 2012

iSCAN: a phoneme-based predictive communication aid for nonspeaking individuals

Ha Trinh; Annalu Waller; Keith Vertanen; Per Ola Kristensson; Vicki L. Hanson

The high incidence of literacy deficits among people with severe speech impairments (SSI) has been well documented. Without literacy skills, people with SSI are unable to effectively use orthographic-based communication systems to generate novel linguistic items in spontaneous conversation. To address this problem, phoneme-based communication systems have been proposed which enable users to create spoken output from phoneme sequences. In this paper, we investigate whether prediction techniques can be employed to improve the usability of such systems. We have developed iSCAN, a phoneme-based predictive communication system, which offers phoneme prediction and phoneme-based word prediction. A pilot study with 16 able-bodied participants showed that our predictive methods led to a 108.4% increase in phoneme entry speed and a 79.0% reduction in phoneme error rate. The benefits of the predictive methods were also demonstrated in a case study with a cerebral palsied participant. Moreover, results of a comparative evaluation conducted with the same participant after 16 sessions using iSCAN indicated that our system outperformed an orthographic-based predictive communication device that the participant has used for over 4 years.


conference of the international speech communication association | 2016

Real-Time Presentation Tracking Using Semantic Keyword Spotting.

Reza Asadi; Harriet J. Fell; Timothy W. Bickmore; Ha Trinh

Given presentation slides with detailed written speaking notes, automatic tracking of oral presentations can help speakers ensure they cover their planned content, and can reduce their anxiety during the speech. Tracking is a more complex problem than speech-to-text alignment, since presenters rarely follow their exact presentation notes, and it must be performed in realtime. In this paper, we propose a novel system that can track the current degree of coverage of each slide’s contents. To do this, the presentation notes for each slide are segmented into sentences, and the words are filtered into keyword candidates. These candidates are then scored based on word specificity and semantic similarity measures to find the most useful keywords for the tracking task. Real-time automatic speech recognition results are matched against the keywords and their synonyms. Sentences are scored based on detected keywords, and the ones with scores higher than a threshold are tagged as covered. We manually and automatically annotated 150 slide presentation recordings to evaluate the system. A simple tracking method, matching speech recognition results against the notes, was used as the baseline. The results show that our approach led to higher accuracy measures compared to the baseline method.


Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies | 2017

RoboCOP: A Robotic Coach for Oral Presentations

Ha Trinh; Reza Asadi; Darren Edge; Timothy W. Bickmore

Rehearsing in front of a live audience is invaluable when preparing for important presentations. However, not all presenters take the opportunity to engage in such rehearsal, due to time constraints, availability of listeners who can provide constructive feedback, or public speaking anxiety. We present RoboCOP, an automated anthropomorphic robot head that acts as a coach to provide spoken feedback during presentation rehearsals at both the individual slide and overall presentation level. The robot offers conversational coaching on three key aspects of presentations: speech quality, content coverage, and audience orientation. The design of the feedback strategies was informed by findings from an exploratory study with academic professionals who were experienced in mentoring students on their presentations. In a within-subjects study comparing RoboCOP to visual feedback and spoken feedback without a robot, the robotic coach was shown to lead to significant improvement in the overall experience of presenters. Results of a second within-subjects evaluation study comparing RoboCOP with existing rehearsal practices show that our system creates a natural, interactive, and motivating rehearsal environment that leads to improved presentation quality.Rehearsing in front of a live audience is invaluable when preparing for important presentations. However, not all presenters take the opportunity to engage in such rehearsal, due to time constraints, availability of listeners who can provide constructive feedback, or public speaking anxiety. We present RoboCOP, an automated anthropomorphic robot head that acts as a coach to provide spoken feedback during presentation rehearsals at both the individual slide and overall presentation level. The robot offers conversational coaching on three key aspects of presentations: speech quality, content coverage, and audience orientation. The design of the feedback strategies was informed by findings from an exploratory study with academic professionals who were experienced in mentoring students on their presentations. In a within-subjects study comparing RoboCOP to visual feedback and spoken feedback without a robot, the robotic coach was shown to lead to significant improvement in the overall experience of presenters. Results of a second within-subjects evaluation study comparing RoboCOP with existing rehearsal practices show that our system creates a natural, interactive, and motivating rehearsal environment that leads to improved presentation quality.


conference on computers and accessibility | 2011

Using a computer intervention to support phonological awareness development of nonspeaking adults

Ha Trinh

The present study investigates the effectiveness of a computer-based intervention to support adults with severe speech and physical impairments (SSPI) in developing their phonological awareness, an essential contributory factor to literacy acquisition. Three participants with SSPI undertook seven intervention sessions during which they were asked to play a training game on an iPad. The game was designed to enable learners to practice their phonological awareness skills independently with minimal instruction from human instructors. Preliminary results of post-intervention assessments demonstrate general positive effects of the intervention upon the phonological awareness and literacy skills of the participants. These results support the use of mainstream technologies to aid learning for individuals with disabilities.


international conference on multimodal interfaces | 2017

IntelliPrompter: speech-based dynamic note display interface for oral presentations

Reza Asadi; Ha Trinh; Harriet J. Fell; Timothy W. Bickmore

The fear of forgetting what to say next is a common problem in oral presentation delivery. To maintain good content coverage and reduce anxiety, many presenters use written notes in conjunction with presentation slides. However, excessive use of notes during delivery can lead to disengagement with the audience and low quality presentations. We designed IntelliPrompter, a speech-based note display system that automatically tracks a presenter’s coverage of each slide’s content and dynamically adjusts the note display interface to highlight the most likely next topic to present. We developed two versions of our intelligent teleprompters using Google Glass and computer screen displays. The design of our system was informed by findings from 36 interviews with presenters and analysis of a presentation note corpus. In a within-subjects study comparing our dynamic screen-based and Google Glass note display interfaces with a static note system, presenters and independent judges expressed a strong preference for the dynamic screen-based system.


intelligent virtual agents | 2016

A Smartphone-Based Virtual Agent for Atrial Fibrillation Education and Counseling

Everlyne Kimani; Timothy W. Bickmore; Ha Trinh; Lazlo Ring; Michael K. Paasche-Orlow; Jared W. Magnani

When deployed on smartphones, virtual agents have the potential to deliver life-saving advice regarding emergency medical conditions, as well as provide a convenient channel for health education to help improve the safety and efficacy of pharmacotherapy. This paper describes the use of a smartphone-based virtual agent that provides counseling to patients with Atrial Fibrillation, along with the results from a pilot acceptance study among patients with the condition. Atrial Fibrillation is a highly prevalent heart rhythm disorder and is known to significantly increase the risk of stroke, heart failure and death. In this study, a virtual agent is deployed in conjunction with a smartphone-based heart rhythm monitor that lets patients obtain real-time diagnostic information on the status of their atrial fibrillation and determine whether immediate action may be needed. The results of the study indicate that participants are satisfied with receiving information about Atrial Fibrillation via the virtual agent.


intelligent virtual agents | 2015

Context-Awareness in a Persistent Hospital Companion Agent

Timothy W. Bickmore; Reza Asadi; Aida Ehyaei; Harriet J. Fell; Lori Henault; Stephen S. Intille; Lisa Quintiliani; Ameneh Shamekhi; Ha Trinh; Katherine Waite; Christopher W. Shanahan; Michael K. Paasche-Orlow

We describe the design and preliminary evaluation of a virtual agent that provides continual bedside companionship and a range of health, information, and entertainment functions to hospital patients during their stay. The agent system uses sensors to enable it to be aware of events in the hospital room and the status of the patient, in order to provide context-sensitive health counseling. Patients in the pilot study responded well to having the agent in their rooms for 1–3 days and engaged in 9.4 conversations per day with the agent on average, using all available functions.


ACM Sigaccess Accessibility and Computing | 2011

Developing a phoneme-based talking joystick for nonspeaking individuals

Ha Trinh

This research investigates the potential of developing a novel phoneme-based assistive communication system for pre-literate individuals with severe speech and physical impairments. Using a force-feedback joystick-like game controller as the access tool, the system enables users to select forty-two phonemes (i.e., English sounds) used in literacy teaching and combine them together to generate spoken messages. What distinguishes this phoneme-based device from other communication systems currently available on the market is that it allows users who have not mastered literacy skills to create novel words and sentences without the need for a visual interface. Natural Language Processing (NLP) technologies, including phoneme-to-speech synthesis, phoneme-based disambiguation and prediction, and haptic force feedback technology are being incorporated into the device to improve its accessibility and usability.

Collaboration


Dive into the Ha Trinh's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Reza Asadi

Northeastern University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lazlo Ring

Northeastern University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Keith Vertanen

Montana Tech of the University of Montana

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Vicki L. Hanson

Rochester Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge