Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Frédéric Apoux is active.

Publication


Featured researches published by Frédéric Apoux.


Journal of the Acoustical Society of America | 2004

Relative importance of temporal information in various frequency regions for consonant identification in quiet and in noise

Frédéric Apoux; Sid P. Bacon

The relative importance of temporal information in broad spectral regions for consonant identification was assessed in normal-hearing listeners. For the purpose of forcing listeners to use primarily temporal-envelope cues, speech sounds were spectrally degraded using four-noise-band vocoder processing Frequency-weighting functions were determined using two methods. The first method consisted of measuring the intelligibility of speech with a hole in the spectrum either in quiet or in noise. The second method consisted of correlating performance with the randomly and independently varied signal-to-noise ratio within each band. Results demonstrated that all bands contributed equally to consonant identification when presented in quiet. In noise, however, both methods indicated that listeners consistently placed relatively more weight upon the highest frequency band. It is proposed that the explanation for the difference in results between quiet and noise relates to the shape of the modulation spectra in adjacent frequency bands. Overall, the results suggest that normal-hearing listeners use a common listening strategy in a given condition. However, this strategy may be influenced by the competing sounds, and thus may vary according to the context. Some implications of the results for cochlear implantees and hearing-impaired listeners are discussed.


Hearing Research | 2009

On the number of auditory filter outputs needed to understand speech: further evidence for auditory channel independence.

Frédéric Apoux; Eric W. Healy

The number of auditory filter outputs required to identify phonemes was estimated in two experiments. Stimuli were divided into 30 contiguous equivalent rectangular bandwidths (ERB(N)) spanning 80-7563Hz. Normal-hearing listeners were presented with limited numbers of bands having frequency locations determined randomly from trial to trial to provide a general view, i.e., irrespective of specific band location, of the number of 1-ERB(N)-wide speech bands needed to identify phonemes. The first experiment demonstrated that 20 such bands are required to accurately identify vowels, and 16 are required to identify consonants. In the second experiment, speech-shaped noise or time-reversed speech was introduced to the non-speech bands at various signal-to-noise ratios. Considerably elevated noise levels were necessary to substantially affect phoneme recognition, confirming a high degree of channel independence in the auditory system. The independence observed between auditory filter outputs supports current views of speech recognition in noise in which listeners extract and combine pieces of information randomly distributed both in time and frequency. These findings also suggest that the ability to partition incoming sounds into a large number of narrow bands, an ability often lost in cases of hearing impairment or cochlear implantation, is critical for speech recognition in noise.


Hearing Research | 2004

Identification of envelope-expanded sentences in normal-hearing and hearing-impaired listeners

Frédéric Apoux; Nicolas Tribut; Xavier Debruille; Christian Lorenzi

The present study examined the effects of temporal-envelope expansion on speech perception. Sentence identification performance was measured in normal-hearing and hearing-impaired listeners for stationary and fluctuating noise. Two expansion schemes were used to increase the depth of the slow (< 16 Hz) and fast (< 256 Hz) amplitude fluctuations of the stimuli. In the first scheme, the envelope of the stimuli was raised to the power 2. In the second scheme, the high- and low-level segments of the envelope were compressed and expanded, respectively. When envelope processing was applied to speech before the addition of noise, the first form of expansion generally degraded identification, while the second form generally improved identification. When envelope processing was applied to speech after the addition of noise, both forms of expansion improved or did not affect identification scores in both groups of listeners when applied to the slowest fluctuations. When applied to the broadest range of fluctuations, both forms of expansion degraded identification. However, in hearing-impaired listeners, the second form of expansion yielded an increase in performance for fluctuating noise. This complex pattern of results will be discussed in light of previous studies on envelope expansion.


Journal of the Acoustical Society of America | 2013

Role and relative contribution of temporal envelope and fine structure cues in sentence recognition by normal-hearing listeners

Frédéric Apoux; Sarah E. Yoho; Carla L. Youngdahl; Eric W. Healy

The present study investigated the role and relative contribution of envelope and temporal fine structure (TFS) to sentence recognition in noise. Target and masker stimuli were added at five different signal-to-noise ratios (SNRs) and filtered into 30 contiguous frequency bands. The envelope and TFS were extracted from each band by Hilbert decomposition. The final stimuli consisted of the envelope of the target/masker sound mixture at x dB SNR and the TFS of the same sound mixture at y dB SNR. A first experiment showed a very limited contribution of TFS cues, indicating that sentence recognition in noise relies almost exclusively on temporal envelope cues. A second experiment showed that replacing the carrier of a sound mixture with noise (vocoder processing) cannot be considered equivalent to disrupting the TFS of the target signal by adding a background noise. Accordingly, a re-evaluation of the vocoder approach as a model to further understand the role of TFS cues in noisy situations may be necessary. Overall, these data are consistent with the view that speech information is primarily extracted from the envelope while TFS cues are primarily used to detect glimpses of the target.


Journal of the Acoustical Society of America | 2008

Differential contribution of envelope fluctuations across frequency to consonant identification in quiet

Frédéric Apoux; Sid P. Bacon

Two experiments investigated the effects of critical bandwidth and frequency region on the use of temporal envelope cues for speech. In both experiments, spectral details were reduced using vocoder processing. In experiment 1, consonant identification scores were measured in a condition for which the cutoff frequency of the envelope extractor was half the critical bandwidth (HCB) of the auditory filters centered on each analysis band. Results showed that performance is similar to those obtained in conditions for which the envelope cutoff was set to 160 Hz or above. Experiment 2 evaluated the impact of setting the cutoff frequency of the envelope extractor to values of 4, 8, and 16 Hz or to HCB in one or two contiguous bands for an eight-band vocoder. The cutoff was set to 16 Hz for all the other bands. Overall, consonant identification was not affected by removing envelope fluctuations above 4 Hz in the low- and high-frequency bands. In contrast, speech intelligibility decreased as the cutoff frequency was decreased in the midfrequency region from 16 to 4 Hz. The behavioral results were fairly consistent with a physical analysis of the stimuli, suggesting that clearly measurable envelope fluctuations cannot be attenuated without affecting speech intelligibility.


Advances in Experimental Medicine and Biology | 2013

A Glimpsing Account of the Role of Temporal Fine Structure Information in Speech Recognition

Frédéric Apoux; Eric W. Healy

Many behavioral studies have reported a significant decrease in intelligibility when the temporal fine structure (TFS) of a sound mixture is replaced with noise or tones (i.e., vocoder processing). This finding has led to the conclusion that TFS information is critical for speech recognition in noise. How the normal -auditory system takes advantage of the original TFS, however, remains unclear. Three -experiments on the role of TFS in noise are described. All three experiments measured speech recognition in various backgrounds while manipulating the envelope, TFS, or both. One experiment tested the hypothesis that vocoder processing may artificially increase the apparent importance of TFS cues. Another experiment evaluated the relative contribution of the target and masker TFS by disturbing only the TFS of the target or that of the masker. Finally, a last experiment evaluated the -relative contribution of envelope and TFS information. In contrast to previous -studies, however, the original envelope and TFS were both preserved - to some extent - in all conditions. Overall, the experiments indicate a limited influence of TFS and suggest that little speech information is extracted from the TFS. Concomitantly, these experiments confirm that most speech information is carried by the temporal envelope in real-world conditions. When interpreted within the framework of the glimpsing model, the results of these experiments suggest that TFS is primarily used as a grouping cue to select the time-frequency regions -corresponding to the target speech signal.


Journal of the Acoustical Society of America | 2011

On the mechanisms involved in the recovery of envelope information from temporal fine structure.

Frédéric Apoux; Rebecca E. Millman; Neal F. Viemeister; Christopher A. Brown; Sid P. Bacon

Three experiments were designed to provide psychophysical evidence for the existence of envelope information in the temporal fine structure (TFS) of stimuli that were originally amplitude modulated (AM). The original stimuli typically consisted of the sum of a sinusoidally AM tone and two unmodulated tones so that the envelope and TFS could be determined a priori. Experiment 1 showed that normal-hearing listeners not only perceive AM when presented with the Hilbert fine structure alone but AM detection thresholds are lower than those observed when presenting the original stimuli. Based on our analysis, envelope recovery resulted from the failure of the decomposition process to remove the spectral components related to the original envelope from the TFS and the introduction of spectral components related to the original envelope, suggesting that frequency- to amplitude-modulation conversion is not necessary to recover envelope information from TFS. Experiment 2 suggested that these spectral components interact in such a way that envelope fluctuations are minimized in the broadband TFS. Experiment 3 demonstrated that the modulation depth at the original carrier frequency is only slightly reduced compared to the depth of the original modulator. It also indicated that envelope recovery is not specific to the Hilbert decomposition.


Journal of the Acoustical Society of America | 2008

Selectivity of modulation interference for consonant identification in normal-hearing listeners

Frédéric Apoux; Sid P. Bacon

The present study sought to establish whether speech recognition can be disrupted by the presence of amplitude modulation (AM) at a remote spectral region, and whether that disruption depends upon the rate of AM. The goal was to determine whether this paradigm could be used to examine which modulation frequencies in the speech envelope are most important for speech recognition. Consonant identification for a band of speech located in either the low- or high-frequency region was measured in the presence of a band of noise located in the opposite frequency region. The noise was either unmodulated or amplitude modulated by a sinusoid, a band of noise with a fixed absolute bandwidth, or a band of noise with a fixed relative bandwidth. The frequency of the modulator was 4, 16, 32, or 64 Hz. Small amounts of modulation interference were observed for all modulator types, irrespective of the location of the speech band. More important, the interference depended on modulation frequency, clearly supporting the existence of selectivity of modulation interference with speech stimuli. Overall, the results suggest a primary role of envelope fluctuations around 4 and 16 Hz without excluding the possibility of a contribution by faster rates.


Journal of the Acoustical Society of America | 2014

Speech-cue transmission by an algorithm to increase consonant recognition in noise for hearing-impaired listeners

Eric W. Healy; Sarah E. Yoho; Yuxuan Wang; Frédéric Apoux; DeLiang Wang

Consonant recognition was assessed following extraction of speech from noise using a more efficient version of the speech-segregation algorithm described in Healy, Yoho, Wang, and Wang [(2013) J. Acoust. Soc. Am. 134, 3029-3038]. Substantial increases in recognition were observed following algorithm processing, which were significantly larger for hearing-impaired (HI) than for normal-hearing (NH) listeners in both speech-shaped noise and babble backgrounds. As observed previously for sentence recognition, older HI listeners having access to the algorithm performed as well or better than young NH listeners in conditions of identical noise. It was also found that the binary masks estimated by the algorithm transmitted speech features to listeners in a fashion highly similar to that of the ideal binary mask (IBM), suggesting that the algorithm is estimating the IBM with substantial accuracy. Further, the speech features associated with voicing, manner of articulation, and place of articulation were all transmitted with relative uniformity and at relatively high levels, indicating that the algorithm and the IBM transmit speech cues without obvious deficiency. Because the current implementation of the algorithm is much more efficient, it should be more amenable to real-time implementation in devices such as hearing aids and cochlear implants.


Journal of the Acoustical Society of America | 2002

Effect of duration on amplitude-modulation masking

Rebecca E. Millman; Christian Lorenzi; Frédéric Apoux; Christian Füllgrabe; Gary G. R. Green; Sid P. Bacon

Detection thresholds for a signal sinusoidal amplitude modulation (SAM) were measured in the presence of an additional, masking SAM. Signal and masker modulations were applied synchronously to a gated white noise carrier. The masker modulation frequency was fixed at 16 Hz and the signal modulation frequency ranged from 2 to 256 Hz. Modulation masking patterns were measured in three listeners for stimulus durations of 500, 1000, and 2000 ms. The results showed that modulation masking was reduced or abolished at signal modulation frequencies of 2, 14, and 18 Hz when stimulus duration was increased from 500 to 2000 ms. The effect at 2 Hz suggests that fringe effects in modulation masking previously reported by Bacon and Grantham [S. P. Bacon and D. W. Grantham, J. Acoust. Soc. Am. 91, 3451–3455 (1992)] were at least partly due to an increase in masker modulation duration. The reduction in modulation masking observed at signal modulation frequencies of 14 and 18 Hz was similar to that at 2 Hz, suggesting that...

Collaboration


Dive into the Frédéric Apoux's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sid P. Bacon

Arizona State University

View shared research outputs
Top Co-Authors

Avatar

Christian Lorenzi

École Normale Supérieure

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge