Nicholas J. Bryan
Stanford University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Nicholas J. Bryan.
international conference on acoustics, speech, and signal processing | 2012
Nicholas J. Bryan; Paris Smaragdis; Gautham J. Mysore
We propose a method to both identify and synchronize multi-camera video recordings within a large collection of video and/or audio files. Landmark-based audio fingerprinting is used to match multiple recordings of the same event together and time-synchronize each file within the groups. Compared to prior work, we offer improvements towards event identification and a new synchronization refinement method that resolves inconsistent estimates and allows non-overlapping content to be synchronized within larger groups of recordings. Furthermore, the audio fingerprinting-based synchronization is shown to be equivalent to an efficient and scalable time-difference-of-arrival method using cross-correlation performed on a non-linearly transformed signal.
human factors in computing systems | 2014
Nicholas J. Bryan; Gautham J. Mysore; Ge Wang
Traditional audio editing tools do not facilitate the task of separating a single mixture recording (e.g. pop song) into its respective sources (e.g. drums, vocal, etc.). Such ability, however, would be very useful for a wide variety of audio applications such as music remixing, audio denoising, and audio-based forensics. To address this issue, we present ISSE - an interactive source separation editor. ISSE is a new open-source, freely available, and cross-platform audio editing tool that enables a user to perform source separation by painting on time-frequency visualizations of sound, resulting in an interactive machine learning system. The system brings to life our previously proposed interaction paradigm and separation algorithm that learns from user-feedback to perform separation. For evaluation, we conducted user studies and compared results between inexperienced and expert users. For a variety of real-world tasks, we found that inexperienced users can achieve good separation quality with minimal instruction and expert users can achieve state-of-the-art separation quality.
international conference on acoustics, speech, and signal processing | 2013
Nicholas J. Bryan; Gautham J. Mysore
We propose an interactive refinement method for supervised and semi-supervised single-channel source separation. The refinement method allows end-users to provide feedback to the separation process by painting on spectrogram displays of intermediate output results. The time-frequency annotations are then used to update the separation estimates and iteratively refine the results. The initial separation is performed using probabilistic latent component analysis and is then extended to incorporate the painting annotations using linear grouping expectation constraints via the framework of posterior regularization. Using a prototype user-interface, we show that the method is able to perform high-quality separation with minimal user-interaction.
new interfaces for musical expression | 2010
Jieun Oh; Jorge Herrera; Nicholas J. Bryan; Luke Dahl; Ge Wang
new interfaces for musical expression | 2010
Nicholas J. Bryan; Jorge Herrera; Jieun Oh; Ge Wang
international computer music conference | 2009
Ge Wang; Nicholas J. Bryan; Jieun Oh; Robert Hamilton
international conference on machine learning | 2013
Nicholas J. Bryan; Gautham J. Mysore
IEEE Transactions on Knowledge and Data Engineering | 2010
Thiago Quirino; Miroslav Kubat; Nicholas J. Bryan
international symposium/conference on music information retrieval | 2013
Nicholas J. Bryan; Gautham J. Mysore; Ge Wang
international symposium/conference on music information retrieval | 2011
Nicholas J. Bryan; Ge Wang