Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Peter Vincent Desouza is active.

Publication


Featured researches published by Peter Vincent Desouza.


international conference on acoustics, speech, and signal processing | 1991

Decision trees for phonological rules in continuous speech

Lalit R. Bahl; Peter Vincent Desouza; Ponani S. Gopalakrishnan; David Nahamoo; Michael Picheny

The authors present an automatic method for modeling phonological variation using decision trees. For each phone they construct a decision tree that specifies the acoustic realization of the phone as a function of the context in which it appears. Several-thousand sentences from a natural language corpus spoken by several speakers are used to construct these decision trees. Experimental results on a 5000-word vocabulary natural language speech recognition task are presented.<<ETX>>


international conference on acoustics, speech, and signal processing | 1991

Automatic phonetic baseform determination

Lalit R. Bahl; Subhro Das; Peter Vincent Desouza; Mark E. Epstein; Robert L. Mercer; Bernard Merialdo; David Nahamoo; Michael Picheny; J. Powell

The authors describe a series of experiments in which the phonetic baseform is deduced automatically for new words by utilizing actual utterances of the new word in conjunction with a set of automatically derived spelling-to-sound rules. Recognition performance was evaluated on new words spoken by two different speakers when the phonetic baseforms were extracted via the above approach. The error rates on these new words were found to be comparable to or better than when the phonetic baseforms were derived by hand, thus validating the basic approach.<<ETX>>


Journal of the Acoustical Society of America | 1989

Determination of phone weights for markov models in a speech recognition system

Lalit R. Bahl; Peter Vincent Desouza; Robert L. Mercer

In a speech recognition system, discrimination between similar-sounding uttered words is improved by weighting the probability vector data stored for the Markov model representing the reference word sequence of phones. The weighting vector is derived for each reference word by comparing similar sounding utterances using Viterbi alignment and multivariate analysis which maximizes the differences between correct and incorrect recognition multivariate distributions.


international conference on acoustics speech and signal processing | 1998

Acoustics-only based automatic phonetic baseform generation

Bhuvana Ramabhadran; Lalit R. Bahl; Peter Vincent Desouza; Mukund Padmanabhan

Phonetic baseforms are the basic recognition units in most speech recognition systems. These baseforms are usually determined by linguists once a vocabulary is chosen and not modified thereafter. However, several applications, such as name dialing, require the user be able to add new words to the vocabulary. These new words are often names, or task-specific jargon, that have user-specific pronunciations. This paper describes a novel method for generating phonetic transcriptions (baseforms) of words based on acoustic evidence alone. It does not require either the spelling or any prior acoustic representation of the new word, is vocabulary independent, and does not have any linguistic constraints (pronunciation rules). Our experiments demonstrate the high decoding accuracies obtained when baseforms deduced using this approach are incorporated into our speech recognizer. Also, the error rates on the added words were found to be comparable to or better than when the baseforms were derived by hand.


international conference on acoustics, speech, and signal processing | 1991

A new class of fenonic Markov word models for large vocabulary continuous speech recognition

Lalit R. Bahl; Jerome R. Bellegarda; Peter Vincent Desouza; Ponani S. Gopalakrishnan; David Nahamoo; Michael Picheny

A technique for constructing hidden Markov models for the acoustic representation of words is described. The models, built from combinations of acoustically based subword units called fenones, are derived automatically from one or more sample utterances of words. They are more flexible than previously reported fenone-based word models and lead to an improved capability of modeling variations in pronunciation. In addition, their construction is simplified, because it can be done using the well-known forward-backward algorithm for the parameter estimation of hidden Markov models. Experimental results obtained on a 5000-word vocabulary continuous speech recognition task are presented to illustrate some of the benefits associated with the new models. Multonic baseforms resulted in a reduction of 16% in the average error rate obtained for ten speakers.<<ETX>>


Computational Linguistics | 1992

Class-based n -gram models of natural language

Peter F. Brown; Peter Vincent Desouza; Robert L. Mercer; Vincent J. Della Pietra; Jenifer C. Lai


Journal of the Acoustical Society of America | 1993

Method and apparatus for the automatic determination of phonological rules as for a continuous speech recognition system

Lalit R. Bahl; Peter F. Brown; Peter Vincent Desouza; Robert L. Mercer


Archive | 1988

Design and construction of a binary-tree system for language modelling

Lalit R. Bahl; Peter F. Brown; Peter Vincent Desouza; Robert L. Mercer


Journal of the Acoustical Society of America | 1990

Constructing Markov model word baseforms from multiple utterances by concatenating model sequences for word segments

Lalit R. Bahl; Peter Vincent Desouza; Robert L. Mercer; Michael Picheny


Archive | 1985

Automatic generation of simple markov model stunted baseforms for words in a vocabulary

Lalit R. Bahl; Peter Vincent Desouza; Robert L. Mercer; Michael Picheny

Researchain Logo
Decentralizing Knowledge