Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Joshua M. Alexander is active.

Publication


Featured researches published by Joshua M. Alexander.


Journal of the Acoustical Society of America | 2004

Informational masking in hearing-impaired and normal-hearing listeners: Sensation level and decision weights

Joshua M. Alexander; Robert A. Lutfi

Informational masking (IM) refers to elevations in signal threshold caused by masker uncertainty. The purpose of this study was to investigate two factors expected to influence IM in hearing-impaired listeners. Masked thresholds for a 2000-Hz signal in the presence of simultaneous multitone maskers were measured in 16 normal-hearing (NH) and 9 hearing-impaired (HI) listeners. The maskers were 70 dB SPL average total power and were comprised of fixed-frequency components between 522 and 8346 Hz that were separated from each other by at least 1/3 oct and from the signal by at least 2/3 octs. Masker uncertainty was manipulated by randomly presenting each masker component with probability p = 0.1,0.2,...,0.9, or 1.0 across different trial blocks. Energetic masking was estimated as the amount of masking for p = 1.0, where masker uncertainty was minimum. IM was estimated as the amount of masking in excess of energetic masking. Decision weights were estimated by a regression of the listeners yes/no responses against the presence or absence of the signal and masker components. The decision weights and sensation levels (SLs) of the stimulus components were incorporated as factors in a model that predicts individual differences in IM based on the level variance (in dB) at the output of independent auditory filters [Lutfi, J. Acoust. Soc. Am. 94, 748-758 (1993)]. The results showed much individual variability in IM for the NH listeners (over 40 dB), but little IM for most HI listeners. When masker components were presented to a group of NH listeners at SLs similar to the HI listeners, IM was also similar to the HI listeners. IM was also similar for both groups when the level per masker component was 10 dB SL. These results suggest that reduced masker SLs for HI listeners decrease IM by effectively reducing masker variance. Weighting efficiencies, computed by comparing each listeners pattern of weights to that of an ideal analytic listener, were a good predictor of individual differences in IM among the NH listeners. For the HI listeners weighting efficiency and IM were unrelated because of the large variation in masker SLs among individual listeners, the small variance in IM, and perhaps because broadened auditory filters in some listeners increased the covariance in auditory filter outputs.


Attention Perception & Psychophysics | 2010

Auditory color constancy: Calibration to reliable spectral properties across nonspeech context and targets

Christian E. Stilp; Joshua M. Alexander; Michael Kiefte; Keith R. Kluender

Brief experience with reliable spectral characteristics of a listening context can markedly alter perception of subsequent speech sounds, and parallels have been drawn between auditory compensation for listening context and visual color constancy. In order to better evaluate such an analogy, the generality of acoustic context effects for sounds with spectral-temporal compositions distinct from speech was investigated. Listeners identified nonspeech sounds—extensively edited samples produced by a French horn and a tenor saxophone—following either resynthesized speech or a short passage of music. Preceding contexts were “colored” by spectral envelope difference filters, which were created to emphasize differences between French horn and saxophone spectra. Listeners were more likely to report hearing a saxophone when the stimulus followed a context filtered to emphasize spectral characteristics of the French horn, and vice versa. Despite clear changes in apparent acoustic source, the auditory system calibrated to relatively predictable spectral characteristics of filtered context, differentially affecting perception of subsequent target nonspeech sounds. This calibration to listening context and relative indifference to acoustic sources operates much like visual color constancy, for which reliable properties of the spectrum of illumination are factored out of perception of color.


Ear and Hearing | 2014

The influence of audibility on speech recognition with nonlinear frequency compression for children and adults with hearing loss.

Ryan W. McCreery; Joshua M. Alexander; Marc A. Brennan; Brenda Hoover; Judy G. Kopun; Patricia G. Stelmachowicz

Objective: The primary goal of nonlinear frequency compression (NFC) and other frequency-lowering strategies is to increase the audibility of high-frequency sounds that are not otherwise audible with conventional hearing aid (HA) processing due to the degree of hearing loss, limited HA bandwidth, or a combination of both factors. The aim of the present study was to compare estimates of speech audibility processed by NFC with improvements in speech recognition for a group of children and adults with high-frequency hearing loss. Design: Monosyllabic word recognition was measured in noise for 24 adults and 12 children with mild to severe sensorineural hearing loss. Stimuli were amplified based on each listener’s audiogram with conventional processing (CP) with amplitude compression or with NFC and presented under headphones using a software-based HA simulator. A modification of the speech intelligibility index (SII) was used to estimate audibility of information in frequency-lowered bands. The mean improvement in SII was compared with the mean improvement in speech recognition. Results: All but 2 listeners experienced improvements in speech recognition with NFC compared with CP, consistent with the small increase in audibility that was estimated using the modification of the SII. Children and adults had similar improvements in speech recognition with NFC. Conclusion: Word recognition with NFC was higher than CP for children and adults with mild to severe hearing loss. The average improvement in speech recognition with NFC (7%) was consistent with the modified SII, which indicated that listeners experienced an increase in audibility with NFC compared with CP. Further studies are necessary to determine whether changes in audibility with NFC are related to speech recognition with NFC for listeners with greater degrees of hearing loss, with a greater variety of compression settings, and using auditory training.


Journal of the Acoustical Society of America | 2008

Spectral tilt change in stop consonant perception

Joshua M. Alexander; Keith R. Kluender

There exists no clear understanding of the importance of spectral tilt for perception of stop consonants. It is hypothesized that spectral tilt may be particularly salient when formant patterns are ambiguous or degraded. Here, it is demonstrated that relative change in spectral tilt over time, not absolute tilt, significantly influences perception of /b/ vs /d/. Experiments consisted of burstless synthesized stimuli that varied in spectral tilt and onset frequency of the second formant. In Experiment 1, tilt of the consonant at voice onset was varied. In Experiment 2, tilt of the vowel steady state was varied. Results of these experiments were complementary and revealed a significant contribution of relative spectral tilt change only when formant information was ambiguous. Experiments 3 and 4 replicated Experiments 1 and 2 in an /aba/-/ada/ context. The additional tilt contrast provided by the initial vowel modestly enhanced effects. In Experiment 5, there was no effect for absolute tilt when consonant and vowel tilts were identical. Consistent with earlier studies demonstrating contrast between successive local spectral features, perceptual effects of gross spectral characteristics are likewise relative. These findings have implications for perception in nonlaboratory environments and for listeners with hearing impairment.


Ear and Hearing | 2014

Effects of frequency compression and frequency transposition on fricative and affricate perception in listeners with normal hearing and mild to moderate hearing loss.

Joshua M. Alexander; Judy G. Kopun; Patricia G. Stelmachowicz

Objectives: The authors have demonstrated that the limited bandwidth associated with conventional hearing aid amplification prevents useful high-frequency speech information from being transmitted. The purpose of this study was to examine the efficacy of two popular frequency-lowering algorithms and one novel algorithm (spectral envelope decimation) in adults with mild to moderate sensorineural hearing loss and in normal-hearing controls. Design: Participants listened monaurally through headphones to recordings of nine fricatives and affricates spoken by three women in a vowel–consonant context. Stimuli were mixed with speech-shaped noise at 10 dB SNR and recorded through a Widex Inteo IN-9 and a Phonak Naída UP V behind-the-ear (BTE) hearing aid. Frequency transposition (FT) is used in the Inteo and nonlinear frequency compression (NFC) used in the Naída. Both devices were programmed to lower frequencies above 4 kHz, but neither device could lower frequencies above 6 to 7 kHz. Each device was tested under four conditions: frequency lowering deactivated (FT-off and NFC-off), frequency lowering activated (FT and NFC), wideband (WB), and a fourth condition unique to each hearing aid. The WB condition was constructed by mixing recordings from the first condition with high-pass filtered versions of the source stimuli. For the Inteo, the fourth condition consisted of recordings made with the same settings as the first, but with the noise-reduction feature activated (FT-off). For the Naída, the fourth condition was the same as the first condition except that source stimuli were preprocessed by a novel frequency compression algorithm, spectral envelope decimation (SED), designed in MATLAB, which allowed for a more complete lowering of the 4 to 10 kHz input band. A follow-up experiment with NFC used Phonak’s Naída SP V BTE, which could also lower a greater range of input frequencies. Results: For normal-hearing and hearing-impaired listeners, performance with FT was significantly worse compared with that in the other conditions. Consistent with previous findings, performance for the hearing-impaired listeners in the WB condition was significantly better than in the FT-off condition. In addition, performance in the SED and WB conditions were both significantly better than in the NFC-off condition and the NFC condition with 6 kHz input bandwidth. There were no significant differences between SED and WB, indicating that improvements in fricative identification obtained by increasing bandwidth can also be obtained using this form of frequency compression. Significant differences between most conditions could be largely attributed to an increase or decrease in confusions for the phonemes /s/ and /z/. In the follow-up experiment, performance in the NFC condition with 10 kHz input bandwidth was significantly better than NFC-off, replicating the results obtained with SED. Furthermore, listeners who performed poorly with NFC-off tended to show the most improvement with NFC. Conclusions: Improvements in the identification of stimuli chosen to be sensitive to the effects of frequency lowering have been demonstrated using two forms of frequency compression (NFC and SED) in individuals with mild to moderate high-frequency sensorineural hearing loss. However, negative results caution against using FT for this population. Results also indicate that the advantage of an extended bandwidth as reported here and elsewhere applies to the input bandwidth for frequency compression (NFC/SED) when the start frequency is ≥4 kHz.


Journal of The American Academy of Audiology | 2014

Paired comparisons of nonlinear frequency compression, extended bandwidth, and restricted bandwidth hearing aid processing for children and adults with hearing loss.

Marc A. Brennan; Ryan W. McCreery; Judy G. Kopun; Brenda Hoover; Joshua M. Alexander; Dawna E. Lewis; Patricia G. Stelmachowicz

BACKGROUND Preference for speech and music processed with nonlinear frequency compression (NFC) and two controls (restricted bandwidth [RBW] and extended bandwidth [EBW] hearing aid processing) was examined in adults and children with hearing loss. PURPOSE The purpose of this study was to determine if stimulus type (music, sentences), age (children, adults), and degree of hearing loss influence listener preference for NFC, RBW, and EBW. RESEARCH DESIGN Design was a within-participant, quasi-experimental study. Using a round-robin procedure, participants listened to amplified stimuli that were (1) frequency lowered using NFC, (2) low-pass filtered at 5 kHz to simulate the RBW of conventional hearing aid processing, or (3) low-pass filtered at 11 kHz to simulate EBW amplification. The examiner and participants were blinded to the type of processing. Using a two-alternative forced-choice task, participants selected the preferred music or sentence passage. STUDY SAMPLE Participants included 16 children (ages 8-16 yr) and 16 adults (ages 19-65 yr) with mild to severe sensorineural hearing loss. INTERVENTION All participants listened to speech and music processed using a hearing aid simulator fit to the Desired Sensation Level algorithm v5.0a. RESULTS Children and adults did not differ in their preferences. For speech, participants preferred EBW to both NFC and RBW. Participants also preferred NFC to RBW. Preference was not related to the degree of hearing loss. For music, listeners did not show a preference. However, participants with greater hearing loss preferred NFC to RBW more than participants with less hearing loss. Conversely, participants with greater hearing loss were less likely to prefer EBW to RBW. CONCLUSIONS Both age groups preferred access to high-frequency sounds, as demonstrated by their preference for either the EBW or NFC conditions over the RBW condition. Preference for EBW can be limited for those with greater degrees of hearing loss, but participants with greater hearing loss may be more likely to prefer NFC. Further investigation using participants with more severe hearing loss may be warranted.


The Senses: A Comprehensive Reference | 2008

Perception of Speech Sounds

Keith R. Kluender; Joshua M. Alexander

Fundamental principles that govern all perception, from hair cell to cortex, are shaping our understanding of the perception of speech and other familiar sounds. Information theory and biology of sensori-neural systems are emphasized in explanations of classic characteristics of speech perception, including perceptual resilience in the face of degradation, signal variability, segmentation, and categorical perception. Multiple experimental findings illustrate how a series of like processes operate upon the acoustic signal with increasing levels of sophistication on the way from waveforms to words. Common to these processes are ways that perceptual systems absorb predictable characteristics of the soundscape, from temporally local to extended periods (learning), and new information (change) is enhanced.


Ear and Hearing | 2015

Effects of WDRC release time and number of channels on output SNR and speech recognition.

Joshua M. Alexander; Katie Masterson

Objectives: The purpose of this study was to investigate the joint effects that wide dynamic range compression (WDRC) release time (RT) and number of channels have on recognition of sentences in the presence of steady and modulated maskers at different signal-to-noise ratios (SNRs). How the different combinations of WDRC parameters affect output SNR and the role this plays in the observed findings were also investigated. Design: Twenty-four listeners with mild to moderate sensorineural hearing loss identified sentences mixed with steady or modulated maskers at three SNRs (−5, 0, and +5 dB) that had been processed using a hearing aid simulator with six combinations of RT (40 and 640 msec) and number of channels (4, 8, and 16). Compression parameters were set using the Desired Sensation Level v5.0a prescriptive fitting method. For each condition, amplified speech and masker levels and the resultant long-term output SNR were measured. Results: Speech recognition with WDRC depended on the combination of RT and number of channels, with the greatest effects observed at 0 dB input SNR, in which mean speech recognition scores varied by 10 to 12% across WDRC manipulations. Overall, effect sizes were generally small. Across both masker types and the three SNRs tested, the best speech recognition was obtained with eight channels, regardless of RT. Increased speech levels, which favor audibility, were associated with the short RT and with an increase in the number of channels. These same conditions also increased masker levels by an even greater amount, for a net decrease in the long-term output SNR. Changes in long-term SNR across WDRC conditions were found to be strongly associated with changes in the temporal envelope shape as quantified by the Envelope Difference Index; however, neither of these factors fully explained the observed differences in speech recognition. Conclusions: A primary finding of this study was that the number of channels had a modest effect when analyzed at each level of RT, with results suggesting that selecting eight channels for a given RT might be the safest choice. Effects were smaller for RT, with results suggesting that short RT was slightly better when only 4 channels were used and that long RT was better when 16 channels were used. Individual differences in how listeners were influenced by audibility, output SNR, temporal distortion, and spectral distortion may have contributed to the size of the effects found in this study. Because only general suppositions could made for how each of these factors may have influenced the overall results of this study, future research would benefit from exploring the predictive value of these and other factors in selecting the processing parameters that maximize speech recognition for individuals.


American Journal of Audiology | 2016

Minimal Hearing Loss: From a Failure-Based Approach to Evidence-Based Practice

Allison M. Winiger; Joshua M. Alexander; Allan O. Diefendorf

PURPOSE A representative sample of the literature on minimal hearing loss (MHL) was reviewed to provide evidence of challenges faced by children with MHL and to establish the need for evidence-based options for early intervention. METHOD Research articles published from 1950 to 2013 were searched in the Medline database using the keywords minimal hearing loss, unilateral hearing loss, and mild hearing loss. References cited in retrieved articles were also reviewed. RESULTS In total, 69 articles contained relevant information about pediatric outcomes and/or intervention for unilateral hearing loss, 50 for mild hearing loss, and 6 for high-frequency hearing loss. Six challenges associated with MHL emerged, and 6 interventions were indicated. Evidence indicates that although some individuals may appear to have no observable speech-language or academic difficulties, others experience considerable difficulties. It also indicates that even though children with MHL may appear to catch up in some areas, difficulties in select domains continue into adulthood. CONCLUSIONS Evidence indicates significant risks associated with untreated MHL. Evidence also demonstrates the need for early intervention and identifies several appropriate intervention strategies; however, no single protocol is appropriate for all children. Therefore, families should be educated about the impact of MHL and about available interventions so that informed decisions can be made.


Journal of the Acoustical Society of America | 2005

Classification and identification of recorded and synthesized impact sounds by practiced listeners, musicians, and nonmusicians

Robert A. Lutfi; Eunmi Oh; Eileen Storm; Joshua M. Alexander

Three experiments were conducted to test the viability of a low-parameter modal model for synthesizing impact sounds to be used in commercial and psychoacoustic research. The model was constrained to have four physically based parameters dictating the amplitude, frequency, and decay of modes. The values of these parameters were selected by ear to roughly match the recordings of ten different resonant objects suspended by hand and struck with different mallets. In experiment 1, neither 35 professional musicians nor 187 college undergraduates could identify which of the two matched sounds was the real recording with better than chance accuracy, though significantly better than chance performance was obtained when modal parameters were selected without the previously imposed physical constraints. In experiment 2, the undergraduates identified the source corresponding to the recorded and synthesized sounds with the same level of accuracy and largely the same pattern of errors. Finally, experiment 3 showed highly practiced listeners to be largely insensitive to changes in the acoustic waveform resulting from an increase in the number of free parameters used in the modal model beyond 3. The results suggest that low-parameter, modal models might be exploited meaningfully in many commercial and research applications involving human perception of impact sounds.

Collaboration


Dive into the Joshua M. Alexander's collaboration.

Top Co-Authors

Avatar

Keith R. Kluender

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar

Robert A. Lutfi

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Eunmi Oh

University of Wisconsin-Madison

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge