Sam Norman-Haignere
Massachusetts Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Sam Norman-Haignere.
The Journal of Neuroscience | 2013
Sam Norman-Haignere; Nancy Kanwisher; Josh H. McDermott
Pitch is a defining perceptual property of many real-world sounds, including music and speech. Classically, theories of pitch perception have differentiated between temporal and spectral cues. These cues are rendered distinct by the frequency resolution of the ear, such that some frequencies produce “resolved” peaks of excitation in the cochlea, whereas others are “unresolved,” providing a pitch cue only via their temporal fluctuations. Despite longstanding interest, the neural structures that process pitch, and their relationship to these cues, have remained controversial. Here, using fMRI in humans, we report the following: (1) consistent with previous reports, all subjects exhibited pitch-sensitive cortical regions that responded substantially more to harmonic tones than frequency-matched noise; (2) the response of these regions was mainly driven by spectrally resolved harmonics, although they also exhibited a weak but consistent response to unresolved harmonics relative to noise; (3) the response of pitch-sensitive regions to a parametric manipulation of resolvability tracked psychophysical discrimination thresholds for the same stimuli; and (4) pitch-sensitive regions were localized to specific tonotopic regions of anterior auditory cortex, extending from a low-frequency region of primary auditory cortex into a more anterior and less frequency-selective region of nonprimary auditory cortex. These results demonstrate that cortical pitch responses are located in a stereotyped region of anterior auditory cortex and are predominantly driven by resolved frequency components in a way that mirrors behavior.
Frontiers in Human Neuroscience | 2010
Nicholas B. Turk-Browne; Sam Norman-Haignere; Gregory McCarthy
Faces activate specific brain regions in fMRI, including the fusiform gyrus (FG) and the posterior superior temporal sulcus (pSTS). The fact that the FG and pSTS are frequently co-activated suggests that they may interact synergistically in a distributed face processing network. Alternatively, the functions implemented by these regions may be encapsulated from each other. It has proven difficult to evaluate these two accounts during visual processing of face stimuli. However, if the FG and pSTS interact during face processing, the substrate for such interactions may be apparent in a correlation of the BOLD timeseries from these two regions during periods of rest when no faces are present. To examine face-specific resting correlations, we developed a new partial functional connectivity approach in which we removed variance from the FG that was shared with other category-selective and control regions. The remaining face-specific FG resting variance was then used to predict resting signals throughout the brain. In two experiments, we observed face-specific resting functional connectivity between FG and pSTS, and importantly, these correlations overlapped precisely with the face-specific pSTS region obtained from independent localizer runs. Additional region-of-interest and pattern analyses confirmed that the FG–pSTS resting correlations were face-specific. These findings support a model in which face processing is distributed among a finite number of connected, but nevertheless face-specialized regions. The discovery of category-specific interactions in the absence of visual input suggests that resting networks may provide a latent foundation for task processing.
Journal of Neurophysiology | 2012
Evelina Fedorenko; Josh H. McDermott; Sam Norman-Haignere; Nancy Kanwisher
Evidence from brain-damaged patients suggests that regions in the temporal lobes, distinct from those engaged in lower-level auditory analysis, process the pitch and rhythmic structure in music. In contrast, neuroimaging studies targeting the representation of music structure have primarily implicated regions in the inferior frontal cortices. Combining individual-subject fMRI analyses with a scrambling method that manipulated musical structure, we provide evidence of brain regions sensitive to musical structure bilaterally in the temporal lobes, thus reconciling the neuroimaging and patient findings. We further show that these regions are sensitive to the scrambling of both pitch and rhythmic structure but are insensitive to high-level linguistic structure. Our results suggest the existence of brain regions with representations of musical structure that are distinct from high-level linguistic representations and lower-level acoustic representations. These regions provide targets for future research investigating possible neural specialization for music or its associated mental processes.
NeuroImage | 2016
Sam Norman-Haignere; Josh H. McDermott
Nonlinearities in the cochlea can introduce audio frequencies that are not present in the sound signal entering the ear. Known as distortion products (DPs), these added frequencies complicate the interpretation of auditory experiments. Sound production systems also introduce distortion via nonlinearities, a particular concern for fMRI research because the Sensimetrics earphones widely used for sound presentation are less linear than most high-end audio devices (due to design constraints). Here we describe the acoustic and neural effects of cochlear and earphone distortion in the context of fMRI studies of pitch perception, and discuss how their effects can be minimized with appropriate stimuli and masking noise. The amplitude of cochlear and Sensimetrics earphone DPs were measured for a large collection of harmonic stimuli to assess effects of level, frequency, and waveform amplitude. Cochlear DP amplitudes were highly sensitive to the absolute frequency of the DP, and were most prominent at frequencies below 300 Hz. Cochlear DPs could thus be effectively masked by low-frequency noise, as expected. Earphone DP amplitudes, in contrast, were highly sensitive to both stimulus and DP frequency (due to prominent resonances in the earphones transfer function), and their levels grew more rapidly with increasing stimulus level than did cochlear DP amplitudes. As a result, earphone DP amplitudes often exceeded those of cochlear DPs. Using fMRI, we found that earphone DPs had a substantial effect on the response of pitch-sensitive cortical regions. In contrast, cochlear DPs had a small effect on cortical fMRI responses that did not reach statistical significance, consistent with their lower amplitudes. Based on these findings, we designed a set of pitch stimuli optimized for identifying pitch-responsive brain regions using fMRI. These stimuli robustly drive pitch-responsive brain regions while producing minimal cochlear and earphone distortion, and will hopefully aid fMRI researchers in avoiding distortion confounds.
Neuron | 2018
Alexander Kell; Daniel Yamins; Erica N. Shook; Sam Norman-Haignere; Josh H. McDermott
A core goal of auditory neuroscience is to build quantitative models that predict cortical responses to natural sounds. Reasoning that a complete model of auditory cortex must solve ecologically relevant tasks, we optimized hierarchical neural networks for speech and music recognition. The best-performing network contained separate music and speech pathways following early shared processing, potentially replicating human cortical organization. The network performed both tasks as well as humans and exhibited human-like errors despite not being optimized to do so, suggesting common constraints on network and human performance. The network predicted fMRI voxel responses substantially better than traditional spectrotemporal filter models throughout auditory cortex. It also provided a quantitative signature of cortical representational hierarchy-primary and non-primary responses were best predicted by intermediate and late network layers, respectively. The results suggest that task optimization provides a powerful set of tools for modeling sensory systems.
bioRxiv | 2018
Sam Norman-Haignere; Josh H. McDermott
A central goal of sensory neuroscience is to construct models that can explain neural responses to complex, natural stimuli. As a consequence, sensory models are often tested by comparing neural responses to natural stimuli with model responses to those stimuli. One challenge is that distinct model features are often correlated across natural stimuli, and thus model features can predict neural responses even if they do not in fact drive them. Here we propose a simple alternative for testing a sensory model: we synthesize stimuli that yield the same model response as a natural stimulus, and test whether the natural and “model-matched” stimulus elicit the same neural response. We used this approach to test whether a common model of auditory cortex – in which spectrogram-like peripheral input is processed by linear spectrotemporal filters – can explain fMRI responses in humans to natural sounds. Prior studies have that shown that this model has good predictive power throughout auditory cortex, but this finding could reflect stimulus-driven correlations. We observed that fMRI voxel responses to natural and model-matched stimuli were nearly equivalent in primary auditory cortex, but that non-primary regions showed highly divergent responses to the two sound sets, suggesting that neurons in non-primary regions extract higher-order properties not made explicit by traditional models. This dissociation between primary and non-primary regions was not clear from model predictions due to the influence of stimulus-driven response correlations. Our methodology enables stronger tests of sensory models and could be broadly applied in other domains. Author Summary Modeling neural responses to natural stimuli is a core goal of sensory neuroscience. Here we propose a new approach for testing sensory models: we synthesize a “model-matched” stimulus that yields the same model response as a natural stimulus, and test whether it produces the same neural response. We used model-matching to test whether a standard model of auditory cortex can explain human cortical responses measured with fMRI. Model-matched stimuli produced nearly equivalent voxel responses in primary auditory cortex, but highly divergent responses in non-primary regions. This dissociation was not evident using more standard approaches for model testing, and suggests that non-primary regions compute higher-order stimulus properties not captured by traditional models. The methodology could be broadly applied in other domains.
Journal of the Acoustical Society of America | 2016
Sam Norman-Haignere; Josh H. McDermott
How does the auditory system recognize instances of the same sound class with distinct acoustic properties? As a case study, we investigated the recognition of environmental sounds at different levels. In principle, level-invariant recognition could be achieved by a normalization mechanism that removes variation in level from listeners’ representation of sound identity. Alternatively, listeners could use level as a cue to aid their recognition, taking advantage of the fact that different sound classes are typically heard at different levels. The latter hypothesis predicts that sounds heard at atypical levels should be more difficult to recognize. We tested this prediction by asking human listeners to identify 300 environmental sounds, each presented at seven different sound levels between 30 and 90 dB SPL. We grouped these 300 sounds into those typically heard at low (e.g. salt-shaker) and high sound levels (e.g., jackhammer) using ratings collected via Mechanical Turk. For typically loud sounds, recognit...
Neuron | 2015
Sam Norman-Haignere; Nancy Kanwisher; Josh H. McDermott
Cerebral Cortex | 2012
Sam Norman-Haignere; Gregory McCarthy; Marvin M. Chun; Nicholas B. Turk-Browne
Journal of Vision | 2015
Alexander Kell; Daniel Yamins; Sam Norman-Haignere; Darren Seibert; Ha Hong; Jim DiCarlo; Josh H. McDermott