Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Chunxi Liu is active.

Publication


Featured researches published by Chunxi Liu.


spoken language technology workshop | 2014

A keyword search system using open source software

Jan Trmal; Guoguo Chen; Daniel Povey; Sanjeev Khudanpur; Pegah Ghahremani; Xiaohui Zhang; Vimal Manohar; Chunxi Liu; Aren Jansen; Dietrich Klakow; David Yarowsky; Florian Metze

Provides an overview of a speech-to-text (STT) and keyword search (KWS) system architecture build primarily on the top of the Kaldi toolkit and expands on a few highlights. The system was developed as a part of the research efforts of the Radical team while participating in the IARPA Babel program. Our aim was to develop a general system pipeline which could be easily and rapidly deployed in any language, independently on the language script and phonological and linguistic features of the language.


international conference on acoustics, speech, and signal processing | 2016

Adapting ASR for under-resourced languages using mismatched transcriptions

Chunxi Liu; Preethi Jyothi; Hao Tang; Vimal Manohar; Rose Sloan; Tyler Kekona; Mark Hasegawa-Johnson; Sanjeev Khudanpur

Mismatched transcriptions of speech in a target language refers to transcriptions provided by people unfamiliar with the language, using English letter sequences. In this work, we demonstrate the value of such transcriptions in building an ASR system for the target language. For different languages, we use less than an hour of mismatched transcriptions to successfully adapt baseline multilingual models built with no access to native transcriptions in the target language. The adapted models provide up to 25% relative improvement in phone error rates on an unseen evaluation set.


IEEE Transactions on Audio, Speech, and Language Processing | 2017

ASR for Under-Resourced Languages From Probabilistic Transcription

Mark Hasegawa-Johnson; Preethi Jyothi; Daniel McCloy; Majid Mirbagheri; Giovanni M. Di Liberto; Amit Das; Bradley Ekin; Chunxi Liu; Vimal Manohar; Hao Tang; Edmund C. Lalor; Nancy F. Chen; Paul Hager; Tyler Kekona; Rose Sloan; Adrian Lee

In many under-resourced languages it is possible to find text, and it is possible to find speech, but transcribed speech suitable for training automatic speech recognition (ASR) is unavailable. In the absence of native transcripts, this paper proposes the use of a probabilistic transcript: A probability mass function over possible phonetic transcripts of the waveform. Three sources of probabilistic transcripts are demonstrated. First, self-training is a well-established semisupervised learning technique, in which a cross-lingual ASR first labels unlabeled speech, and is then adapted using the same labels. Second, mismatched crowdsourcing is a recent technique in which nonspeakers of the language are asked to write what they hear, and their nonsense transcripts are decoded using noisy channel models of second-language speech perception. Third, EEG distribution coding is a new technique in which nonspeakers of the language listen to it, and their electrocortical response signals are interpreted to indicate probabilities. ASR was trained in four languages without native transcripts. Adaptation using mismatched crowdsourcing significantly outperformed self-training, and both significantly outperformed a cross-lingual baseline. Both EEG distribution coding and text-derived phone language models were shown to improve the quality of probabilistic transcripts derived from mismatched crowdsourcing.


international conference on acoustics, speech, and signal processing | 2017

An empirical evaluation of zero resource acoustic unit discovery

Chunxi Liu; Jinyi Yang; Ming Sun; Santosh Kesiraju; Alena Rott; Lucas Ondel; Pegah Ghahremani; Najim Dehak; Lukas Burget; Sanjeev Khudanpur

Acoustic unit discovery (AUD) is a process of automatically identifying a categorical acoustic unit inventory from speech and producing corresponding acoustic unit tokenizations. AUD provides an important avenue for unsupervised acoustic model training in a zero resource setting where expert-provided linguistic knowledge and transcribed speech are unavailable. Therefore, to further facilitate zero-resource AUD process, in this paper, we demonstrate acoustic feature representations can be significantly improved by (i) performing linear discriminant analysis (LDA) in an unsupervised self-trained fashion, and (ii) leveraging resources of other languages through building a multilingual bottleneck (BN) feature extractor to give effective cross-lingual generalization. Moreover, we perform comprehensive evaluations of AUD efficacy on multiple downstream speech applications, and their correlated performance suggests that AUD evaluations are feasible using different alternative language resources when only a subset of these evaluation resources can be available in typical zero resource applications.


international conference on acoustics, speech, and signal processing | 2016

Context-dependent point process models for keyword search and detection-based ASR

Chunxi Liu; Aren Jansen; Sanjeev Khudanpur

The point process model (PPM) for keyword search (KWS) is a whole-word parametric approach that characterizes each query type by the timing of phonetic events observed during its production. In this paper, we first extend the PPM modeling framework to operate on context-dependent phonetic event patterns instead of monophone patterns considered in the past, which provides significant KWS improvements. Second, we use the context-dependent PPMs to drive a detection-based speech recognition architecture thats runs parallel word detectors covering the whole vocabulary and uses the independent detections to construct lattices that can be used for both KWS indexing and LVCSR decoding. This strategy produces significant improvements over the original PPM KWS framework and provides an encouraging first attempt at PPM-based LVCSR.


conference of the international speech communication association | 2014

Low-resource open vocabulary keyword search using point process models.

Chunxi Liu; Aren Jansen; Guoguo Chen; Keith Kintzley; Jan Trmal; Sanjeev Khudanpur


conference of the international speech communication association | 2018

Automatic Speech Recognition and Topic Identification from Speech for Almost-Zero-Resource Languages.

Matthew Wiesner; Chunxi Liu; Lucas Ondel; Craig Harman; Vimal Manohar; Jan Trmal; Zhongqiang Huang; Najim Dehak; Sanjeev Khudanpur


arXiv: Computation and Language | 2018

Low-Resource Contextual Topic Identification on Speech.

Chunxi Liu; Matthew Wiesner; Shinji Watanabe; Craig Harman; Jan Trmal; Najim Dehak; Sanjeev Khudanpur


arXiv: Computation and Language | 2018

Automatic Speech Recognition and Topic Identification for Almost-Zero-Resource Languages.

Matthew Wiesner; Chunxi Liu; Lucas Ondel; Craig Harman; Vimal Manohar; Jan Trmal; Zhongqiang Huang; Najim Dehak; Sanjeev Khudanpur


arXiv: Computation and Language | 2018

The JHU Speech LOREHLT 2017 System: Cross-Language Transfer for Situation-Frame Detection.

Matthew Wiesner; Chunxi Liu; Lucas Ondel; Craig Harman; Vimal Manohar; Jan Trmal; Zhongqiang Huang; Sanjeev Khudanpur; Najim Dehak

Collaboration


Dive into the Chunxi Liu's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jan Trmal

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar

Vimal Manohar

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

Najim Dehak

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lucas Ondel

Brno University of Technology

View shared research outputs
Top Co-Authors

Avatar

Aren Jansen

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar

Guoguo Chen

Johns Hopkins University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge