Mostafa Ali Shahin
Texas A&M University at Qatar
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Mostafa Ali Shahin.
spoken language technology workshop | 2012
Mostafa Ali Shahin; Beena Ahmed; Kirrie J. Ballard
Technology based speech therapy systems are severely handicapped due to the absence of accurate prosodic event identification algorithms. This paper introduces an automatic method for the classification of strong-weak (SW) and weak-strong (WS) stress patterns in children speech with American English accent, for use in the assessment of the speech dysprosody. We investigate the ability of two sets of features used to train classifiers to identify the variation in lexical stress between two consecutive syllables. The first set consists of traditional features derived from measurements of pitch, intensity and duration, whereas the second set consists of energies of different filter banks. Three different classifiers were used in the experiments: an Artificial Neural Network (ANN) classifier with a single hidden layer, Support Vector Machine (SVM) classifier with both linear and Gaussian kernels and the Maximum Entropy modeling (MaxEnt). these features. Best results were obtained using an ANN classifier and a combination of the two sets of features. The system correctly classified 94% of the SW stress patterns and 76% of the WS stress patterns.
conference on computers and accessibility | 2013
Avinash Parnandi; Virendra Karappa; Youngpyo Son; Mostafa Ali Shahin; Jacqueline McKechnie; Kirrie J. Ballard; Beena Ahmed; Ricardo Gutierrez-Osuna
We present a multi-tier system for the remote administration of speech therapy to children with apraxia of speech. The system uses a client-server architecture model and facilitates task-oriented remote therapeutic training in both in-home and clinical settings. Namely, the system allows a speech therapist to remotely assign speech production exercises to each child through a web interface, and the child to practice these exercises on a mobile device. The mobile app records the childs utterances and streams them to a back-end server for automated scoring by a speech-analysis engine. The therapist can then review the individual recordings and the automated scores through a web interface, provide feedback to the child, and adapt the training program as needed. We validated the system through a pilot study with children diagnosed with apraxia of speech, and their parents and speech therapists. Here we describe the overall client-server architecture, middleware tools used to build the system, the speech-analysis tools for automatic scoring of recorded utterances, and results from the pilot study. Our results support the feasibility of the system as a complement to traditional face-to-face therapy through the use of mobile tools and automated speech analysis algorithms.
ACM Transactions on Accessible Computing | 2015
Avinash Parnandi; Virendra Karappa; Tian Lan; Mostafa Ali Shahin; Jacqueline McKechnie; Kirrie J. Ballard; Beena Ahmed; Ricardo Gutierrez-Osuna
We present a multitier system for the remote administration of speech therapy to children with apraxia of speech. The system uses a client-server architecture model and facilitates task-oriented remote therapeutic training in both in-home and clinical settings. The system allows a speech language pathologist (SLP) to remotely assign speech production exercises to each child through a web interface and the child to practice these exercises in the form of a game on a mobile device. The mobile app records the childs utterances and streams them to a back-end server for automated scoring by a speech-analysis engine. The SLP can then review the individual recordings and the automated scores through a web interface, provide feedback to the child, and adapt the training program as needed. We have validated the system through a pilot study with children diagnosed with apraxia of speech, their parents, and SLPs. Here, we describe the overall client-server architecture, middleware tools used to build the system, speech-analysis tools for automatic scoring of utterances, and present results from a clinical study. Our results support the feasibility of the system as a complement to traditional face-to-face therapy through the use of mobile tools and automated speech analysis algorithms.
spoken language technology workshop | 2014
Mostafa Ali Shahin; Beena Ahmed; Kirrie J. Ballard
Lexical stress is a key diagnostic marker of disordered speech as it strongly affects speech perception. In this paper we introduce an automated method to classify between the different lexical stress patterns in childrens speech. A deep neural network is used to classify between strong-weak (SW), weak-strong (WS) and equal-stress (SS/WW) patterns in English by measuring the articulation change between the two successive syllables. The deep neural network architecture is trained using a set of acoustic features derived from pitch, duration and intensity measurements along with the energies in different frequency bands. We compared the performance of the deep neural classifier to a traditional single hidden layer MLP. Results show that the deep neural classifier outperforms the traditional MLP. The accuracy of the deep neural system is approximately 85% when classifying between the unequal stress patterns (SW/WS) and greater than 70% when classifying both equal and unequal stress patterns.
international conference on acoustics, speech, and signal processing | 2016
Mostafa Ali Shahin; Ricardo Gutierrez-Osuna; Beena Ahmed
Technology-based therapy tools can be of great benefit to children with developmental speech disabilities as they typically require sustained practice with a speech therapist for several years. Towards this aim, over the past 4 years we have developed speech processing tools to automatically detect common errors in disordered speech. This paper presents an automated technique to identify incorrect lexical stress. Specifically, we describe a deep neural network (DNN) that can be used to classify the four different bisyllabic stress patterns: strong-weak (SW), weak-strong (WS), strong-strong (SS) and weak-weak (WW). We derive input features for the DNN from the duration, pitch, intensity and spectral energy on each of the two consecutive syllables. Using these features, we achieve 93% correct classification between SW/WS stress patterns and 88% correct classification of the four bisyllabic patterns on speech from typically developing children, while we obtain 73.4% classification between SW/WS in disordered speech. These figures represent a two-fold reduction in error rates compared to our prior work, which used a DNN with differential features from consecutive syllables.
international conference of the ieee engineering in medicine and biology society | 2017
Lamana Mulaffer; Mostafa Ali Shahin; Martin Glos; Thomas Penzel; Beena Ahmed
Sleep disorders are becoming increasingly prevalent in society. However most of the burgeoning research on automated sleep analysis has been in the realm of sleep stage classification with limited focus on accurately diagnosing these disorders. In this paper, we explore two different models to discriminate between control and insomnia patients using support vector machine (SVM) classifiers. We validated the models using data collected from 124 participants, 70 control and 54 with insomnia. The first model uses 57 features derived from two channels of EEG data and achieved an accuracy of 81%. The second model uses 15 features from each participants hypnogram and achieved an accuracy of 74%. The impetus behind using these two models is to follow the clinicians diagnostic decision-making process where both the EEG signals and the hypnograms are used. These results demonstrate that there is potential for further experimentation and improvement of the predictive capability of the models to help in diagnosing sleep disorders like insomnia.
conference of the international speech communication association | 2016
Mostafa Ali Shahin; Julien Epps; Beena Ahmed
Prosodic features are important for the intelligibility and proficiency of stress-timed languages such as English and Arabic. Producing the appropriate lexical stress is challenging for second language (L2) learners, in particular, those whose first language (L1) is a syllable-timed language such as Spanish, French, etc. In this paper we introduce a method for automatic classification of lexical stress to be integrated into computer-aided pronunciation learning (CAPL) tools for L2 learning. We trained two different deep learning architectures, the deep feedforward neural network (DNN) and the deep convolutional neural network (CNN) using a set of temporal and spectral features related to the intensity, duration, pitch and energies in different frequency bands. The system was applied on both English (kids and adult) and Arabic (adult) speech corpora collected from native speakers. Our method results in error rates of 9%, 7% and 18% when tested on the English children corpus, English adult corpus and Arabic adult corpus respectively.
conference of the international speech communication association | 2006
Sherif M. Abdou; Salah Eldeen Hamid; Mohsen A. Rashwan; Abdurrahman Samir; Ossama Abdel-Hamid; Mostafa Ali Shahin; Waleed Nazih
language resources and evaluation | 2006
Mustafa Yaseen; Mohammed Attia; Bente Maegaard; Khalid Choukri; Niklas Paulsson; S. Haamid; Steven Krauwer; Chomicha Bendahman; Hanne Fersøe; Mohsen A. Rashwan; Bassam Haddad; Chafic Mokbel; A. Mouradi; A. Al-Kufaishi; Mostafa Ali Shahin; Noureddine Chenfour; Ahmed Ragheb
Speech Communication | 2015
Mostafa Ali Shahin; Beena Ahmed; Avinash Parnandi; Virendra Karappa; Jacqueline McKechnie; Kirrie J. Ballard; Ricardo Gutierrez-Osuna