Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Jim Kannampuzha is active.

Publication


Featured researches published by Jim Kannampuzha.


non linear speech processing | 2009

Towards a neurocomputational model of speech production and perception

Bernd J. Kröger; Jim Kannampuzha; Christiane Neuschaefer-Rube

The limitation in performance of current speech synthesis and speech recognition systems may result from the fact that these systems are not designed with respect to the human neural processes of speech production and perception. A neurocomputational model of speech production and perception is introduced which is organized with respect to human neural processes of speech production and perception. The production-perception model comprises an artificial computer-implemented vocal tract as a front-end module, which is capable of generating articulatory speech movements and acoustic speech signals. The structure of the production-perception model comprises motor and sensory processing pathways. Speech knowledge is collected during training stages which imitate early stages of speech acquisition. This knowledge is stored in artificial self-organizing maps. The current neurocomputational model is capable of producing and perceiving vowels, VC-, and CV-syllables (V=vowels and C=voiced plosives). Basic features of natural speech production and perception are predicted from this model in a straight forward way: Production of speech items is feedforward and feedback controlled and phoneme realizations vary within perceptually defined regions. Perception is less categorical in the case of vowels in comparison to consonants. Due to its human-like production-perception processing the model should be discussed as a basic module for more technical relevant approaches for high-quality speech synthesis and for high performance speech recognition.


Cognitive Computation | 2011

Movements and Holds in Fluent Sentence Production of American Sign Language: The Action-Based Approach

Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Emily Kaufmann; Irene Mittelberg

The importance of bodily movements in the production and perception of communicative actions has been shown for the spoken language modality and accounted for by a theory of communicative actions (Cogn. Process. 2010;11:187–205). In this study, the theory of communicative actions was adapted to the sign language modality; we tested the hypothesis that in the fluent production of short sign language sentences, strong-hand manual sign actions are continuously ongoing without holds, while co-manual oral expression actions (i.e. sign-related actions of the lips, jaw, and tip of the tongue) and co-manual facial expression actions (i.e. actions of the eyebrows, eyelids, etc.), as well as weak-hand actions, show considerable holds. An American Sign Language (ASL) corpus of 100 sentences was analyzed by visually inspecting each frame-to-frame difference (30 frames/s) for separating movement and hold phases for each manual, oral, and facial action. Excluding fingerspelling and signs in sentence-final position, no manual holds were found for the strong hand (0%; the weak hand is not considered), while oral holds occurred in 22% of all oral expression actions and facial holds occurred for all facial expression actions analyzed (100%). These results support the idea that in each language modality, the dominant articulatory system (vocal tract or manual system) determines the timing of actions. In signed languages, in which manual actions are dominant, holds occur mainly in co-manual oral and co-manual facial actions. Conversely, in spoken language, vocal tract actions (i.e. actions of the lips, tongue, jaw, velum, and vocal folds) are dominant; holds occur primarily in co-verbal manual and co-verbal facial actions.


COST'11 Proceedings of the 2011 international conference on Cognitive Behavioural Systems | 2011

The neurophonetic model of speech processing ACT: structure, knowledge acquisition, and function modes

Bernd J. Kröger; Jim Kannampuzha; Cornelia Eckers; Stefan Heim; Emily Kaufmann; Christiane Neuschaefer-Rube

Speech production and speech perception are important human capabilities comprising cognitive as well as sensorimotor functions. This paper summarizes our work developing a neurophonetic model for speech processing, called ACT, which was carried out over the last seven years. The function modes of the model are production, perception, and acquisition. The name of our model reflects the fact that vocal tract ACTions, which constitute motor plans of speech items, are the central units in this model. Specifically (i) the structure of the model, (ii) the acquired knowledge, and (iii) the correspondence between the models structure and specific brain regions are discussed.


COST'09 Proceedings of the Second international conference on Development of Multimodal Interfaces: active Listening and Synchrony | 2009

Gesture duration and articulator velocity in plosive-vowel-transitions

Dominik Bauer; Jim Kannampuzha; Phil Hoole; Bernd J. Kröger

In this study the gesture duration and articulator velocity in con-sonant-vowel-transitions has been analysed using electromagnetic articulography (EMA). The receiver coils where placed on the tongue, lips and teeth. We found onset and offset durations which are statistically significant for a special articulator. The duration of the offset is affected by the degree of opening of the following vowel. The acquired data is intended to tune the control model of an articulatory speech synthesizer to improve the acoustic quality of plosive-vowel-transitions.


EPJ Nonlinear Biomedical Physics | 2014

Associative learning and self-organization as basic principles for simulating speech acquisition, speech production, and speech perception

Bernd J. Kröger; Jim Kannampuzha; Emily Kaufmann


conference of the international speech communication association | 2006

Modeling sensory-to-motor mappings using neural nets and a 3d articulatory speech synthesizer.

Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Christiane Neuschaefer-Rube


Archive | 2009

Phonetotopy within a neurocomputational model of speech production and speech acquisition

Bernd J. Kröger; Jim Kannampuzha; Anja Lowit; Christiane Neuschaefer-Rube; S. Fuchs; H. Loevenbruck; D. Pape; P. Perrier


Cross-Modal Analysis of Speech, Gestures, Gaze and Facial Expressions | 2009

Articulatory Speech Re-synthesis: Profiting from Natural Acoustic Speech Data

Dominik Bauer; Jim Kannampuzha; Bernd J. Kröger


AVSP | 2008

A Neurofunctional Model of Speech Production Including Aspects of Auditory and Audio-Visual Speech Perception

Bernd J. Kröger; Jim Kannampuzha


Proceedings of the Third COST 2102 international training school conference on Toward autonomous, adaptive, and context-aware multimodal interfaces: theoretical and practical issues | 2010

Categorical perception of consonants and vowels: evidence from a neurophonetic model of speech production and perception

Bernd J. Kröger; Peter Birkholz; Jim Kannampuzha; Christiane Neuschaefer-Rube

Collaboration


Dive into the Jim Kannampuzha's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Peter Birkholz

Dresden University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hermann Ney

RWTH Aachen University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Stefan Heim

RWTH Aachen University

View shared research outputs
Top Co-Authors

Avatar

Anja Lowit

University of Strathclyde

View shared research outputs
Researchain Logo
Decentralizing Knowledge