Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Erik Bresch is active.

Publication


Featured researches published by Erik Bresch.


Journal of the Acoustical Society of America | 2006

Synchronized and noise-robust audio recordings during realtime magnetic resonance imaging scans

Erik Bresch; Jon F. Nielsen; Krishna S. Nayak; Shrikanth Narayanan

This letter describes a data acquisition setup for recording, and processing, running speech from a person in a magnetic resonance imaging (MRI) scanner. The main focus is on ensuring synchronicity between image and audio acquisition, and in obtaining good signal to noise ratio to facilitate further speech analysis and modeling. A field-programmable gate array based hardware design for synchronizing the scanner image acquisition to other external data such as audio is described. The audio setup itself features two fiber optical microphones and a noise-canceling filter. Two noise cancellation methods are described including a novel approach using a pulse sequence specific model of the gradient noise of the MRI scanner. The setup is useful for scientific speech production studies. Sample results of speech and singing data acquired and processed using the proposed method are given.


IEEE Transactions on Medical Imaging | 2009

Region Segmentation in the Frequency Domain Applied to Upper Airway Real-Time Magnetic Resonance Images

Erik Bresch; Shrikanth Narayanan

We describe a method for unsupervised region segmentation of an image using its spatial frequency domain representation. The algorithm was designed to process large sequences of real-time magnetic resonance (MR) images containing the 2-D midsagittal view of a human vocal tract airway. The segmentation algorithm uses an anatomically informed object model, whose fit to the observed image data is hierarchically optimized using a gradient descent procedure. The goal of the algorithm is to automatically extract the time-varying vocal tract outline and the position of the articulators to facilitate the study of the shaping of the vocal tract during speech production.


Journal of Phonetics | 2009

Timing effects of syllable structure and stress on nasals: a real-time MRI examination.

Dani Byrd; Stephen Tobin; Erik Bresch; Shrikanth Narayanan

The coordination of velum and oral gestures for English [n] is studied using real-time MRI movies to reconstruct vocal tract aperture functions. This technique allows for the examination of parts of the vocal tract otherwise inaccessible to dynamic imaging or movement tracking. The present experiment considers syllable onset, coda, and juncture geminate nasals and also addresses the effects of a variety of word stress patterns on segment internal coordination. We find a bimodal timing pattern in which near-synchrony of velum lowering and tongue tip raising characterizes the timing for onsets and temporal lag between the gestures is characteristic for codas, supporting and extending the findings of Krakow (1989), 1993) for [m]. Intervocalic word-internal nasals are found to have timing patterns that are sensitive to the local stress context, which suggests the presence of an underlying timing specification that can yield flexibly. We consider these findings in light of the gestural coupling structures described by Goldstein and colleagues (Goldstein, Byrd, & Saltzman 2006; Nam, Goldstein, and Saltzman in press; Goldstein, Nam, Saltzman, & Chitoran 2008).


IEEE Signal Processing Magazine | 2008

Seeing speech: Capturing vocal tract shaping using real-time magnetic resonance imaging [Exploratory DSP]

Erik Bresch; Yoon-Chul Kim; Krishna S. Nayak; Dani Byrd; Shrikanth Narayanan

In this paper real-time (RT) magnetic resonance imaging (MRI) is used to study speech production especially capturing vocal tract shaping.


Journal of the Acoustical Society of America | 2014

Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (TC)

Shrikanth Narayanan; Asterios Toutios; Vikram Ramanarayanan; Adam C. Lammert; Jangwon Kim; Sungbok Lee; Krishna S. Nayak; Yoon Chul Kim; Yinghua Zhu; Louis Goldstein; Dani Byrd; Erik Bresch; Athanasios Katsamanis; Michael Proctor

USC-TIMIT is an extensive database of multimodal speech production data, developed to complement existing resources available to the speech research community and with the intention of being continuously refined and augmented. The database currently includes real-time magnetic resonance imaging data from five male and five female speakers of American English. Electromagnetic articulography data have also been presently collected from four of these speakers. The two modalities were recorded in two independent sessions while the subjects produced the same 460 sentence corpus used previously in the MOCHA-TIMIT database. In both cases the audio signal was recorded and synchronized with the articulatory data. The database and companion software are freely available to the research community.


Journal of the Acoustical Society of America | 2013

Paralinguistic mechanisms of production in human “beatboxing”: A real-time magnetic resonance imaging study

Michael Proctor; Erik Bresch; Dani Byrd; Krishna S. Nayak; Shrikanth Narayanan

Real-time magnetic resonance imaging (rtMRI) was used to examine mechanisms of sound production by an American male beatbox artist. rtMRI was found to be a useful modality with which to study this form of sound production, providing a global dynamic view of the midsagittal vocal tract at frame rates sufficient to observe the movement and coordination of critical articulators. The subjects repertoire included percussion elements generated using a wide range of articulatory and airstream mechanisms. Many of the same mechanisms observed in human speech production were exploited for musical effect, including patterns of articulation that do not occur in the phonologies of the artists native languages: ejectives and clicks. The data offer insights into the paralinguistic use of phonetic primitives and the ways in which they are coordinated in this style of musical performance. A unified formalism for describing both musical and phonetic dimensions of human vocal percussion performance is proposed. Audio and video data illustrating production and orchestration of beatboxing sound effects are provided in a companion annotated corpus.


Journal of the Acoustical Society of America | 2009

Analysis of pausing behavior in spontaneous speech using real-time magnetic resonance imaging of articulation

Vikram Ramanarayanan; Erik Bresch; Dani Byrd; Louis Goldstein; Shrikanth Narayanan

It is hypothesized that pauses at major syntactic boundaries (i.e., grammatical pauses), but not ungrammatical (e.g., word search) pauses, are planned by a high-level cognitive mechanism that also controls the rate of articulation around these junctures. Real-time magnetic resonance imaging is used to analyze articulation at and around grammatical and ungrammatical pauses in spontaneous speech. Measures quantifying the speed of articulators were developed and applied during these pauses as well as during their immediate neighborhoods. Grammatical pauses were found to have an appreciable drop in speed at the pause itself as compared to ungrammatical pauses, which is consistent with our hypothesis that grammatical pauses are indeed choreographed by a central cognitive planner.


Journal of the Acoustical Society of America | 2010

Real-time magnetic resonance imaging investigation of resonance tuning in soprano singing.

Erik Bresch; Shrikanth Narayanan

This article investigates using real-time magnetic resonance imaging the vocal tract shaping of 5 soprano singers during the production of two-octave scales of sung vowels. A systematic shift of the first vocal tract resonance frequency with respect to the fundamental is shown to exist for high vowels across all subjects. No consistent systematic effect on the vocal tract resonance could be shown across all of the subjects for other vowels or for the second vocal tract resonance.


Anesthesia & Analgesia | 2017

Calibration of Contactless Pulse Oximetry.

Wim Verkruysse; Marek Janusz Bartula; Erik Bresch; Mukul Julius Rocque; Mohammed Meftah; Ihor Olehovych Kirenko

BACKGROUND: Contactless, camera-based photoplethysmography (PPG) interrogates shallower skin layers than conventional contact probes, either transmissive or reflective. This raises questions on the calibratability of camera-based pulse oximetry. METHODS: We made video recordings of the foreheads of 41 healthy adults at 660 and 840 nm, and remote PPG signals were extracted. Subjects were in normoxic, hypoxic, and low temperature conditions. Ratio-of-ratios were compared to reference SpO2 from 4 contact probes. RESULTS: A calibration curve based on artifact-free data was determined for a population of 26 individuals. For an SpO2 range of approximately 83% to 100% and discarding short-term errors, a root mean square error of 1.15% was found with an upper 99% one-sided confidence limit of 1.65%. Under normoxic conditions, a decrease in ambient temperature from 23 to 7°C resulted in a calibration error of 0.1% (±1.3%, 99% confidence interval) based on measurements for 3 subjects. PPG signal strengths varied strongly among individuals from about 0.9 × 10−3 to 4.6 × 10−3 for the infrared wavelength. CONCLUSIONS: For healthy adults, the results present strong evidence that camera-based contactless pulse oximetry is fundamentally feasible because long-term (eg, 10 minutes) error stemming from variation among individuals expressed as A*rms is significantly lower (<1.65%) than that required by the International Organization for Standardization standard (<4%) with the notion that short-term errors should be added. A first illustration of such errors has been provided with A**rms = 2.54% for 40 individuals, including 6 with dark skin. Low signal strength and subject motion present critical challenges that will have to be addressed to make camera-based pulse oximetry practically feasible.


Journal of the Acoustical Society of America | 2006

Syllable structure effects on velum‐oral coordination evaluated with real‐time MRI

Stephen Tobin; Dani Byrd; Erik Bresch; Shrikanth Narayanan

Seminal work by Krakow (PhD. Diss., Yale University, 1989, Phonetics and Phonology 5, 87–116, 1993) has indicated that intrasegmental coordination of articulatory gestures varies systematically as a function of syllable position. Specifically, onset and coda nasals display different lag times between the velum and oral gestural components. In onset nasals, velum lowering is achieved roughly synchronously with oral constriction, while in codas, velum lowering precedes oral constriction. However, Krakow’s study was limited by its reliance on indirect measures of velum movement. Nor did her work examine nasals other than [m] or syllable organizations other than singleton onsets and codas. Our study employs real‐time magnetic resonance imaging [Narayanan et al., J. Acoust. Soc. Am. 115, 1771–1776 (2004); Mathiak et al., Int. J. of Lang. & Comm. Disorders 35, 419–425 (2000)] to acquire direct information about the key vocal tract variables of velum aperture and oral constriction, thereby allowing the calculati...

Collaboration


Dive into the Erik Bresch's collaboration.

Top Co-Authors

Avatar

Shrikanth Narayanan

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Dani Byrd

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Louis Goldstein

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Krishna S. Nayak

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Sungbok Lee

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Vikram Ramanarayanan

University of Southern California

View shared research outputs
Researchain Logo
Decentralizing Knowledge