Dominic Telaar
Karlsruhe Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Dominic Telaar.
Frontiers in Neuroscience | 2015
Christian Herff; Dominic Heger; Adriana de Pesters; Dominic Telaar; Peter Brunner; Tanja Schultz
It has long been speculated whether communication between humans and machines based on natural speech related cortical activity is possible. Over the past decade, studies have suggested that it is feasible to recognize isolated aspects of speech from neural signals, such as auditory features, phones or one of a few isolated words. However, until now it remained an unsolved challenge to decode continuously spoken speech from the neural substrate associated with speech and language processing. Here, we show for the first time that continuously spoken speech can be decoded into the expressed words from intracranial electrocorticographic (ECoG) recordings.Specifically, we implemented a system, which we call Brain-To-Text that models single phones, employs techniques from automatic speech recognition (ASR), and thereby transforms brain activity while speaking into the corresponding textual representation. Our results demonstrate that our system can achieve word error rates as low as 25% and phone error rates below 50%. Additionally, our approach contributes to the current understanding of the neural basis of continuous speech production by identifying those cortical regions that hold substantial information about individual phones. In conclusion, the Brain-To-Text system described in this paper represents an important step toward human-machine communication based on imagined speech.
international conference on acoustics, speech, and signal processing | 2012
Ngoc Thang Vu; Dau-Cheng Lyu; Jochen Weiner; Dominic Telaar; Tim Schlippe; Fabian Blaicher; Eng Siong Chng; Tanja Schultz; Haizhou Li
This paper presents first steps toward a large vocabulary continuous speech recognition system (LVCSR) for conversational Mandarin-English code-switching (CS) speech. We applied state-of-the-art techniques such as speaker adaptive and discriminative training to build the first baseline system on the SEAME corpus [1] (South East Asia Mandarin-English). For acoustic modeling, we applied different phone merging approaches based on the International Phonetic Alphabet (IPA) and Bhattacharyya distance in combination with discriminative training to improve accuracy. On language model level, we investigated statistical machine translation (SMT) - based text generation approaches for building code-switching language models. Furthermore, we integrated the provided information from a language identification system (LID) into the decoding process by using a multi-stream approach. Our best 2-pass system achieves a Mixed Error Rate (MER) of 36.6% on the SEAME development set.
IEEE Transactions on Audio, Speech, and Language Processing | 2015
Heike Adel; Ngoc Thang Vu; Katrin Kirchhoff; Dominic Telaar; Tanja Schultz
This paper presents our latest investigations on different features for factored language models for Code-Switching speech and their effect on automatic speech recognition (ASR) performance. We focus on syntactic and semantic features which can be extracted from Code-Switching text data and integrate them into factored language models. Different possible factors, such as words, part-of-speech tags, Brown word clusters, open class words and clusters of open class word embeddings are explored. The experimental results reveal that Brown word clusters, part-of-speech tags and open-class words are the most effective at reducing the perplexity of factored language models on the Mandarin-English Code-Switching corpus SEAME. In ASR experiments, the model containing Brown word clusters and part-of-speech tags and the model also including clusters of open class word embeddings yield the best mixed error rate results. In summary, the best language model can significantly reduce the perplexity on the SEAME evaluation set by up to 10.8% relative and the mixed error rate by up to 3.4% relative.
The third International Workshop on Spoken Languages Technologies for Under-resourced Languages | 2012
Jochen Weiner; Ngoc Thang Vu; Dominic Telaar; Florian Metze; Tanja Schultz; Dau-Cheng Lyu; Eng Siong Chng; Haizhou Li
conference of the international speech communication association | 2014
Dominic Telaar; Michael Wand; Dirk Gehrig; Felix Putze; Christoph Amma; Dominic Heger; Ngoc Thang Vu; Mark Erhardt; Tim Schlippe; Matthias Janke; Christian Herff; Tanja Schultz
SLTU | 2014
Heike Adel; Katrin Kirchhoff; Dominic Telaar; Ngoc Thang Vu; Tim Schlippe; Tanja Schultz
conference of the international speech communication association | 2015
Dominic Heger; Christian Herff; Adriana de Pesters; Dominic Telaar; Peter Brunner; Tanja Schultz
conference of the international speech communication association | 2014
Heike Adel; Katrin Kirchhoff; Ngoc Thang Vu; Dominic Telaar; Tanja Schultz
language resources and evaluation | 2016
Jochen Weiner; Claudia Frankenberg; Dominic Telaar; Britta Wendelstein; Johannes Schröder; Tanja Schultz
conference of the international speech communication association | 2014
Heike Adel; Dominic Telaar; Ngoc Thang Vu; Katrin Kirchhoff; Tanja Schultz