Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Dawen Liang is active.

Publication


Featured researches published by Dawen Liang.


international conference on acoustics, speech, and signal processing | 2015

Speech dereverberation using a learned speech model

Dawen Liang; Matthew D. Hoffman; Gautham J. Mysore

We present a general single-channel speech dereverberation method based on an explicit generative model of reverberant and noisy speech. To regularize the model, we use a pre-learned speech model of clean and dry speech as a prior and perform posterior inference over the latent clean speech. The reverberation kernel and additive noise are estimated under the maximum-likelihood framework. Our model assumes no prior knowledge about specific speakers or rooms, and consequently our method can automatically adapt to various reverberant and noisy conditions. We evaluate the proposed model with both simulated data and real recordings from the REVERB Challenge1 in the task of speech enhancement and obtain results comparable to or better than the state-of-the-art.


Computer Music Journal | 2014

Active scores: Representation and synchronization in human-computer performance of popular music

Roger B. Dannenberg; Nicolas Gold; Dawen Liang; Guangyu Xia

Computers have the potential to significantly extend the practice of popular music based on steady tempo and mostly determined form. There are significant challenges to overcome, however, due to constraints including accurate timing based on beats and adherence to a form or structure despite possible changes that may occur, possibly even during performance. We describe an approach to synchronization across media that takes into account latency due to communication delays and audio buffering. We also address the problem of mapping from a conventional score with repeats and other structures to an actual performance, which can involve both “flattening” the score and rearranging it, as is common in popular music. Finally, we illustrate the possibilities of the score as a bidirectional user interface in a real-time system for music performance, allowing the user to direct the computer through a digitally displayed score, and allowing the computer to indicate score position back to human performers.


international conference on acoustics, speech, and signal processing | 2014

Speech decoloration based on the product-of-filters model

Dawen Liang; Daniel P. W. Ellis; Matthew D. Hoffman; Gautham J. Mysore

We present a single-channel speech decoloration method based on a recently proposed generative product-of-filters (PoF) model. We take a spectral approach and attempt to learn the magnitude response of the actual coloration filter, given only the degraded speech signal. Experiments on synthetic data demonstrate that the proposed method effectively captures both coarse and fine structure of the coloration filter. On real recordings, we find that simply subtracting the learned coloration filter from the log-spectra yields promising decoloration results.


Computer Music Journal | 2014

Methods and prospects for human-computer performance of popular music

Roger B. Dannenberg; Nicolas Gold; Dawen Liang; Guangyu Xia

Computers are often used in performance of popular music, but most often in very restricted ways, such as keyboard synthesizers where musicians are in complete control, or pre-recorded or sequenced music where musicians follow the computers drums or click track. An interesting and yet little-explored possibility is the computer as highly autonomous performer of popular music, capable of joining a mixed ensemble of computers and humans. Considering the skills and functional requirements of musicians leads to a number of predictions about future human–computer music performance (HCMP) systems for popular music. We describe a general architecture for such systems and describe some early implementations and our experience with them.


Proceedings of the 14th Python in Science Conference | 2015

librosa: Audio and Music Signal Analysis in Python

Brian McFeek; Colin Raffel; Dawen Liang; Matt McVicar; Eric Battenberg; Oriol Nieto


international symposium/conference on music information retrieval | 2015

Content-Aware Collaborative Music Recommendation Using Pre-trained Neural Networks.

Dawen Liang; Minshu Zhan; Daniel P. W. Ellis


international symposium/conference on music information retrieval | 2014

MIR_EVAL: A Transparent Implementation of Common MIR Metrics.

Colin Raffel; Brian McFee; Eric J. Humphrey; Justin Salamon; Oriol Nieto; Dawen Liang; Daniel P. W. Ellis


international symposium/conference on music information retrieval | 2013

Beta Process Sparse Nonnegative Matrix Factorization for Music.

Dawen Liang; Matthew D. Hoffman; Daniel P. W. Ellis


international symposium/conference on music information retrieval | 2014

Codebook-based Scalable Music Tagging with Poisson Matrix Factorization.

Dawen Liang; John Paisley; Daniel P. W. Ellis


new interfaces for musical expression | 2011

A Framework for Coordination and Synchronization of Media.

Dawen Liang; Guangyu Xia; Roger B. Dannenberg

Collaboration


Dive into the Dawen Liang's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brian McFee

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Matt McVicar

National Institute of Advanced Industrial Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Guangyu Xia

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge