Sheila Flanagan
University of Cambridge
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Sheila Flanagan.
Journal of the Acoustical Society of America | 2016
Tudor-Cătălin Zorilă; Yannis Stylianou; Sheila Flanagan; Brian C. J. Moore
A model for the loudness of time-varying sounds [Glasberg and Moore (2012). J. Audio. Eng. Soc. 50, 331-342] was assessed for its ability to predict the loudness of sentences that were processed to either decrease or increase their dynamic fluctuations. In a paired-comparison task, subjects compared the loudness of unprocessed and processed sentences that had been equalized in (1) root-mean square (RMS) level; (2) the peak long-term loudness predicted by the model; (3) the mean long-term loudness predicted by the model. Method 2 was most effective in equating the loudness of the original and processed sentences.
Journal of the Acoustical Society of America | 2017
Tudor-Cătălin Zorilă; Yannis Stylianou; Sheila Flanagan; Brian C. J. Moore
Four algorithms designed to enhance the intelligibility of speech when noise is added after processing were evaluated under the constraint that the speech should have the same loudness before and after processing, as determined using a loudness model. The algorithms applied spectral modifications and two of them included dynamic-range compression. On average, the methods with dynamic-range compression required the least level adjustment to equate loudness for the unprocessed and processed speech. Subjects with normal-hearing (experiment 1) and mild-to-moderate hearing loss (experiment 2) were tested using unmodified and enhanced speech presented in speech-shaped noise (SSN) and a competing speaker (CS). The results showed (a) the algorithms with dynamic-range compression yielded the largest intelligibility gains in both experiments and for both types of background; (b) the algorithms without dynamic-range compression either yielded benefit only with the SSN or yielded no consistent benefit; (c) speech reception thresholds for unprocessed speech were higher for hearing-impaired than for normal-hearing subjects, by about 2 dB for the SSN and 6 dB for the CS. It is concluded that the enhancement methods incorporating dynamic-range compression can improve intelligibility under the equal-loudness constraint for both normal-hearing and hearing-impaired subjects and for both steady and fluctuating backgrounds.
Journal of the Acoustical Society of America | 2018
Sheila Flanagan; Usha Goswami
Recent models of the neural encoding of speech suggest a core role for amplitude modulation (AM) structure, particularly regarding AM phase alignment. Accordingly, speech tasks that measure linguistic development in children may exhibit systematic properties regarding AM structure. Here, the acoustic structure of spoken items in child phonological and morphological tasks, phoneme deletion and plural elicitation, was investigated. The phase synchronisation index (PSI), reflecting the degree of phase alignment between pairs of AMs, was computed for 3 AM bands (delta, theta, beta/low gamma; 0.9-2.5 Hz, 2.5-12 Hz, 12-40 Hz, respectively), for five spectral bands covering 100-7250 Hz. For phoneme deletion, data from 94 child participants with and without dyslexia was used to relate AM structure to behavioural performance. Results revealed that a significant change in magnitude of the phase synchronisation index (ΔPSI) of slower AMs (delta-theta) systematically accompanied both phoneme deletion and plural elicitation. Further, children with dyslexia made more linguistic errors as the delta-theta ΔPSI increased. Accordingly, ΔPSI between slower temporal modulations in the speech signal systematically distinguished test items from accurate responses and predicted task performance. This may suggest that sensitivity to slower AM information in speech is a core aspect of phonological and morphological development.
Trends in hearing | 2018
Sheila Flanagan; Tudor-Cătălin Zorilă; Yannis Stylianou; Brian C. J. Moore
Auditory processing disorder (APD) may be diagnosed when a child has listening difficulties but has normal audiometric thresholds. For adults with normal hearing and with mild-to-moderate hearing impairment, an algorithm called spectral shaping with dynamic range compression (SSDRC) has been shown to increase the intelligibility of speech when background noise is added after the processing. Here, we assessed the effect of such processing using 8 children with APD and 10 age-matched control children. The loudness of the processed and unprocessed sentences was matched using a loudness model. The task was to repeat back sentences produced by a female speaker when presented with either speech-shaped noise (SSN) or a male competing speaker (CS) at two signal-to-background ratios (SBRs). Speech identification was significantly better with SSDRC processing than without, for both groups. The benefit of SSDRC processing was greater for the SSN than for the CS background. For the SSN, scores were similar for the two groups at both SBRs. For the CS, the APD group performed significantly more poorly than the control group. The overall improvement produced by SSDRC processing could be useful for enhancing communication in a classroom where the teacher’s voice is broadcast using a wireless system.
PLOS ONE | 2018
João Pessoa Araújo; Sheila Flanagan; Alexandre Castro-Caldas; Usha Goswami
The temporal modulation structure of speech plays a key role in neural encoding of the speech signal. Amplitude modulations (AMs, quasi-rhythmic changes in signal energy or intensity) in speech are encoded by neuronal oscillations (rhythmic variations in neural excitability in large cell networks) that oscillate at matching temporal rates. To date, however, all neural studies have investigated adult-directed speech (ADS) as produced and perceived by highly literate adults. Whether temporal features of ADS vary with the skills of the speaker, for example literacy skills, is currently unknown. Here we analyse the temporal structure of ADS spoken by illiterate, low literate (≤ 4 years of literacy) and highly literate (≥ 12 years of literacy) adults. We find that illiterates produce speech differently. Spontaneous conversational speech produced by illiterate adults showed significantly less synchronised coupling between AM bands (less phase synchronisation) than conversational speech produced by low literate and highly literate adults, and contained significantly fewer syllables per second. There was also a significant relationship between years of literacy and the amount of theta-band energy in conversational speech. When asked to produce rhythmic proverbs learned in childhood, all groups could produce speech with similar AM phase synchronisation, suggesting that the differences in spontaneous conversational speech were not caused by physiological constraints. The data suggest that the temporal modulation structure of spoken language changes with the acquisition of cultural skills like literacy that are usually a product of schooling. There is a cultural effect on the temporal modulation structure of spoken language.
conference of the international speech communication association | 2016
Tudor-Catalin Zorila; Sheila Flanagan; Brian C. J. Moore; Yannis Stylianou
Most recently proposed near-end speech enhancement methods have been evaluated with the overall power (RMS) of the speech held constant. While significant intelligibility gains have been reported in various noisy conditions, an equal-RMS constraint may lead to enhancement solutions that increase the loudness of the original speech. Comparable effects might be produced simply by increasing the power of the original speech, which also leads to an increase in loudness. Here we suggest modifying the equal-RMS constraint to one of equal loudness between the original and the modified signals, based on a loudness model for time-varying sounds. Four state-of-the-art speechin-noise intelligibility enhancement systems were evaluated under the equal-loudness constraint, using intelligibility tests with normal-hearing listeners. Results were compared with those obtained under the equal-RMS constraint. The methods based on spectral shaping and dynamic range compression yielded significant intelligibility gains regardless of the constraint, while for the method without dynamic range compression the intelligibility gain was lower under the equal-loudness than under the equal-RMS constraint.
Journal of The Audio Engineering Society | 2005
Sheila Flanagan; Brian C. J. Moore; Michael A. Stone
Journal of The Audio Engineering Society | 1998
Neil Harris; Sheila Flanagan; Malcolm J. Hawksford
Journal of The Audio Engineering Society | 1998
Neil Harris; Sheila Flanagan; Malcolm J. Hawksford
Jaro-journal of The Association for Research in Otolaryngology | 2017
Robert P. Carlyon; Sheila Flanagan; John M. Deeks