Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yun-hsuan Sung is active.

Publication


Featured researches published by Yun-hsuan Sung.


international conference on acoustics, speech, and signal processing | 2011

Recognizing English queries in Mandarin Voice Search

Hung-An Chang; Yun-hsuan Sung; Brian Strope; Francoise Beaufays

Recent improvements in speech recognition technology, along with increased computing power and bigger datasets, have considerably improved the state of the art in the field, making it possible for commercial apps such as Google Voice Search to serve users in their everyday mobile search needs. Deploying such systems in various countries has shown us the extent to which multilingualism is present in some cultures, and the need for better solutions to handle it in our speech recognition systems. In this paper, we describe a few early data sharing and model combination experiments we did to improve the recognition of English queries made to Mandarin Voice Search, in Taiwan. We obtained a 12% relative sentence accuracy improvement over a baseline system already including some support for English queries.


international conference on acoustics, speech, and signal processing | 2009

Revisiting graphemes with increasing amounts of data

Yun-hsuan Sung; Thad Hughes; Françoise Beaufays; Brian Strope

Letter units, or graphemes, have been reported in the literature as a surprisingly effective substitute to the more traditional phoneme units, at least in languages that enjoy a strong correspondence between pronunciation and orthography. For English however, where letter symbols have less acoustic consistency, previously reported results fell short of systems using highly-tuned pronunciation lexicons. Grapheme units simplify system design, but since graphemes map to a wider set of acoustic realizations than phonemes, we should expect grapheme-based acoustic models to require more training data to capture these variations. In this paper, we compare the rate of improvement of grapheme and phoneme systems trained with datasets ranging from 450 to 1200 hours of speech. We consider various grapheme unit configurations, including using letter-specific, onset, and coda units. We show that the grapheme systems improve faster and, depending on the lexicon, reach or surpass the phoneme baselines with the largest training set.


international conference on acoustics, speech, and signal processing | 2012

Recognition of multilingual speech in mobile applications

Hui Lin; Jui-ting Huang; Francoise Beaufays; Brian Strope; Yun-hsuan Sung

We evaluate different architectures to recognize multilingual speech for real-time mobile applications. In particular, we show that combining the results of several recognizers greatly outperforms other solutions such as training a single large multilingual system or using an explicit language identification system to select the appropriate recognizer. Experiments are conducted on a trilingual English-French-Mandarin mobile speech task. The data set includes Google searches, Maps queries, as well as more general inputs such as email and short message dictation. Without pre-specifying the input language, the combined system achieves comparable accuracy to that of the monolingual systems when the input language is known. The combined system is also roughly 5% absolute better than an explicit language identification approach, and 10% better than a single large multilingual system.


Archive | 2012

Recognizing speech in multiple languages

Yun-hsuan Sung; Francoise Beaufays; Brian Strope; Hui Lin; Jui-ting Huang


arXiv: Computation and Language | 2016

Conversational Contextual Cues: The Case of Personalization and History for Response Ranking

Rami Al-Rfou; Marc Pickett; Javier Snaider; Yun-hsuan Sung; Brian Strope


Archive | 2012

Recognizing different versions of a language

Francoise Beaufays; Brian Strope; Yun-hsuan Sung


arXiv: Computation and Language | 2018

Universal Sentence Encoder

Daniel M. Cer; Yinfei Yang; Sheng-yi Kong; Nan Hua; Nicole Lyn Untalan Limtiaco; Rhomni St. John; Noah Constant; Mario Guajardo-Céspedes; Steve Yuan; Chris Tar; Yun-hsuan Sung; Brian Strope; Ray Kurzweil


arXiv: Computation and Language | 2017

Efficient Natural Language Response Suggestion for Smart Reply

Matthew Henderson; Rami Al-Rfou; Brian Strope; Yun-hsuan Sung; László Lukács; Ruiqi Guo; Sanjiv Kumar; Balint Miklos; Ray Kurzweil


conference of the international speech communication association | 2013

Written-domain language modeling for automatic speech recognition.

Hasim Sak; Yun-hsuan Sung; Francoise Beaufays; Cyril Allauzen


Archive | 2013

Written-domain language modeling with decomposition

Hasim Sak; Yun-hsuan Sung; Cyril Allauzen

Collaboration


Dive into the Yun-hsuan Sung's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge