Jim Kannampuzha
RWTH Aachen University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Jim Kannampuzha.
non linear speech processing | 2009
Bernd J. Kröger; Jim Kannampuzha; Christiane Neuschaefer-Rube
The limitation in performance of current speech synthesis and speech recognition systems may result from the fact that these systems are not designed with respect to the human neural processes of speech production and perception. A neurocomputational model of speech production and perception is introduced which is organized with respect to human neural processes of speech production and perception. The production-perception model comprises an artificial computer-implemented vocal tract as a front-end module, which is capable of generating articulatory speech movements and acoustic speech signals. The structure of the production-perception model comprises motor and sensory processing pathways. Speech knowledge is collected during training stages which imitate early stages of speech acquisition. This knowledge is stored in artificial self-organizing maps. The current neurocomputational model is capable of producing and perceiving vowels, VC-, and CV-syllables (V=vowels and C=voiced plosives). Basic features of natural speech production and perception are predicted from this model in a straight forward way: Production of speech items is feedforward and feedback controlled and phoneme realizations vary within perceptually defined regions. Perception is less categorical in the case of vowels in comparison to consonants. Due to its human-like production-perception processing the model should be discussed as a basic module for more technical relevant approaches for high-quality speech synthesis and for high performance speech recognition.
Cognitive Computation | 2011
Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Emily Kaufmann; Irene Mittelberg
The importance of bodily movements in the production and perception of communicative actions has been shown for the spoken language modality and accounted for by a theory of communicative actions (Cogn. Process. 2010;11:187–205). In this study, the theory of communicative actions was adapted to the sign language modality; we tested the hypothesis that in the fluent production of short sign language sentences, strong-hand manual sign actions are continuously ongoing without holds, while co-manual oral expression actions (i.e. sign-related actions of the lips, jaw, and tip of the tongue) and co-manual facial expression actions (i.e. actions of the eyebrows, eyelids, etc.), as well as weak-hand actions, show considerable holds. An American Sign Language (ASL) corpus of 100 sentences was analyzed by visually inspecting each frame-to-frame difference (30 frames/s) for separating movement and hold phases for each manual, oral, and facial action. Excluding fingerspelling and signs in sentence-final position, no manual holds were found for the strong hand (0%; the weak hand is not considered), while oral holds occurred in 22% of all oral expression actions and facial holds occurred for all facial expression actions analyzed (100%). These results support the idea that in each language modality, the dominant articulatory system (vocal tract or manual system) determines the timing of actions. In signed languages, in which manual actions are dominant, holds occur mainly in co-manual oral and co-manual facial actions. Conversely, in spoken language, vocal tract actions (i.e. actions of the lips, tongue, jaw, velum, and vocal folds) are dominant; holds occur primarily in co-verbal manual and co-verbal facial actions.
COST'11 Proceedings of the 2011 international conference on Cognitive Behavioural Systems | 2011
Bernd J. Kröger; Jim Kannampuzha; Cornelia Eckers; Stefan Heim; Emily Kaufmann; Christiane Neuschaefer-Rube
Speech production and speech perception are important human capabilities comprising cognitive as well as sensorimotor functions. This paper summarizes our work developing a neurophonetic model for speech processing, called ACT, which was carried out over the last seven years. The function modes of the model are production, perception, and acquisition. The name of our model reflects the fact that vocal tract ACTions, which constitute motor plans of speech items, are the central units in this model. Specifically (i) the structure of the model, (ii) the acquired knowledge, and (iii) the correspondence between the models structure and specific brain regions are discussed.
COST'09 Proceedings of the Second international conference on Development of Multimodal Interfaces: active Listening and Synchrony | 2009
Dominik Bauer; Jim Kannampuzha; Phil Hoole; Bernd J. Kröger
In this study the gesture duration and articulator velocity in con-sonant-vowel-transitions has been analysed using electromagnetic articulography (EMA). The receiver coils where placed on the tongue, lips and teeth. We found onset and offset durations which are statistically significant for a special articulator. The duration of the offset is affected by the degree of opening of the following vowel. The acquired data is intended to tune the control model of an articulatory speech synthesizer to improve the acoustic quality of plosive-vowel-transitions.
EPJ Nonlinear Biomedical Physics | 2014
Bernd J. Kröger; Jim Kannampuzha; Emily Kaufmann
conference of the international speech communication association | 2006
Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Christiane Neuschaefer-Rube
Archive | 2009
Bernd J. Kröger; Jim Kannampuzha; Anja Lowit; Christiane Neuschaefer-Rube; S. Fuchs; H. Loevenbruck; D. Pape; P. Perrier
Cross-Modal Analysis of Speech, Gestures, Gaze and Facial Expressions | 2009
Dominik Bauer; Jim Kannampuzha; Bernd J. Kröger
AVSP | 2008
Bernd J. Kröger; Jim Kannampuzha
Proceedings of the Third COST 2102 international training school conference on Toward autonomous, adaptive, and context-aware multimodal interfaces: theoretical and practical issues | 2010
Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Christiane Neuschaefer-Rube