Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Frederick Jelinek is active.

Publication


Featured researches published by Frederick Jelinek.


IEEE Transactions on Information Theory | 1974

Optimal decoding of linear codes for minimizing symbol error rate (Corresp.)

Lalit R. Bahl; John Cocke; Frederick Jelinek; Josef Raviv

The general problem of estimating the a posteriori probabilities of the states and transitions of a Markov source observed through a discrete memoryless channel is considered. The decoding of linear block and convolutional codes to minimize symbol error probability is shown to be a special case of this problem. An optimal decoding algorithm is derived.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 1983

A Maximum Likelihood Approach to Continuous Speech Recognition

Lalit R. Bahl; Frederick Jelinek; Robert L. Mercer

Speech recognition is formulated as a problem of maximum likelihood decoding. This formulation requires statistical models of the speech production process. In this paper, we describe a number of statistical models for use in speech recognition. We give special attention to determining the parameters for such models from sparse data. We also describe two decoding methods, one appropriate for constrained artificial languages and one appropriate for more realistic decoding tasks. To illustrate the usefulness of the methods described, we review a number of decoding results that have been obtained with them.


human language technology | 1991

Procedure for quantitatively comparing the syntactic coverage of English grammars

Steven P. Abney; S. Flickenger; Claudia Gdaniec; C. Grishman; Philip Harrison; Donald Hindle; Robert Ingria; Frederick Jelinek; Judith L. Klavans; Mark Liberman; Mitchell P. Marcus; Salim Roukos; Beatrice Santorini; Tomek Strzalkowski; Ezra Black

The problem of quantitatively comparing the performance of different broad-coverage grammars of English has to date resisted solution. Prima facie, known English grammars appear to disagree strongly with each other as to the elements of even the simplest sentences. For instance, the grammars of Steve Abney (Bellcore), Ezra Black (IBM), Dan Flickinger (Hewlett Packard), Claudia Gdaniec (Logos), Ralph Grishman and Tomek Strzalkowski (NYU), Phil Harrison (Boeing), Don Hindle (AT&T), Bob Ingria (BBN), and Mitch Marcus (U. of Pennsylvania) recognize in common only the following constituents, when each grammarian provides the single parse which he/she would ideally want his/her grammar to specify for three sample Brown Corpus sentences:The famed Yankee Clipper, now retired, has been assisting (as (a batting coach)).One of those capital-gains ventures, in fact, has saddled him (with Gore Court).He said this constituted a (very serious) misuse (of the (Criminal court) processes).


Computer Speech & Language | 2000

Structured language modeling

Ciprian Chelba; Frederick Jelinek

This paper presents an attempt at using the syntactic structure in natural language for improved language models for speech recognition. The structured language model merges techniques in automatic parsing and language modeling using an original probabilistic parameterization of a shift-reduce parser. A maximum likelihood re-estimation procedure belonging to the class of expectation-maximization algorithms is employed for training the model. Experiments on the Wall Street Journal and Switchboard corpora show improvement in both perplexity and word error rate?word lattice rescoring?over the standard 3-gram language model.


IEEE Transactions on Information Theory | 1975

Design of a linguistic statistical decoder for the recognition of continuous speech

Frederick Jelinek; Lalit R. Bahl; Robert L. Mercer

Most current attempts at automatic speech recognition are formulated in an artificial intelligence framework. In this paper we approach the problem from an information-theoretic point of view. We describe the overall structure of a linguistic statistical decoder (LSD) for the recognition of continuous speech. The input to the decoder is a string of phonetic symbols estimated by an acoustic processor (AP). For each phonetic string, the decoder finds the most likely input sentence. The decoder consists of four major subparts: 1) a statistical model of the language being recognized; 2) a phonemic dictionary and statistical phonological rules characterizing the speaker; 3) a phonetic matching algorithm that computes the similarity between phonetic strings, using the performance characteristics of the AP; 4) a word level search control. The details of each of the subparts and their interaction during the decoding process are discussed.


international conference on computational linguistics | 1988

A statistical approach to language translation

Peter F. Brown; John Cocke; S. Della Pietra; V. Della Pietra; Frederick Jelinek; Robert Leroy Mercer; Paul S. Roossin

An approach to automatic translation is outlined that utilizes techniques of statistical information extraction from large data bases. The method is based on the availability of pairs of large corresponding texts that are translations of each other. In our case, the texts are in English and French.Fundamental to the technique is a complex glossary of correspondence of fixed locutions. The steps of the proposed translation process are: (1) Partition the source text into a set of fixed locutions. (2) Use the glossary plus contextual information to select the corresponding set of fixed locutions into a sequence forming the target sentence. (3) Arrange the words of the target fixed locutions into a sequence forming the target sentence.We have developed statistical techniques facilitating both the automatic creation of the glossary, and the performance of the three translation steps, all on the basis of an alignment of corresponding sentences in the two texts.While we are not yet able to provide examples of French / English translation, we present some encouraging intermediate results concerning glossary creation and the arrangement of target word sequences.


meeting of the association for computational linguistics | 1993

Towards History-based Grammars: Using Richer Models for Probabilistic Parsing

Ezra Black; Frederick Jelinek; John Lafrerty; David M. Magerman; Robert L. Mercer; Salim Roukos

We describe a generative probabilistic model of natural language, which we call HBG, that takes advantage of detailed linguistic information to resolve ambiguity. HBG incorporates lexical, syntactic, semantic, and structural information from the parse tree into the disambiguation process in a novel way. We use a corpus of bracketed sentences, called a Treebank, in combination with decision tree building to tease out the relevant aspects of a parse tree that will determine the correct parse of a sentence. This stands in contrast to the usual approach of further grammar tailoring via the usual linguistic introspection in the hope of generating the correct parse. In head-to-head tests against one of the best existing robust probabilistic parsing models, which we call P-CFG, the HBG model significantly outperforms P-CFG, increasing the parsing accuracy rate from 60% to 75%, a 37% reduction in error.


meeting of the association for computational linguistics | 1998

Exploiting Syntactic Structure for Language Modeling

Ciprian Chelba; Frederick Jelinek

The paper presents a language model that develops syntactic structure and uses it to extract meaningful information from the word history, thus enabling the use of long distance dependencies. The model assigns probability to every joint sequence of words-binary-parse-structure with headword annotation and operates in a left-to-right manner --- therefore usable for automatic speech recognition. The model, its probabilistic parameterization, and a set of experiments meant to evaluate its predictive power are presented; an improvement over standard trigram modeling is achieved.


IEEE Transactions on Information Theory | 1975

Decoding for channels with insertions, deletions, and substitutions with applications to speech recognition

Lalit R. Bahl; Frederick Jelinek

A model for channels in which an input sequence can produce output sequences of varying length is described. An efficient computational procedure for calculating Pr \{Y\mid X\} is devised, where X = x_1,x_2,\cdots,x_M and Y = y_1,y_2,\cdots,y_N are the input and output of the channel. A stack decoding algorithm for decoding on such channels is presented. The appropriate likelihood function is derived. Channels with memory are considered. Some applications to speech and character recognition are discussed.


Archive | 1992

Basic Methods of Probabilistic Context Free Grammars

Frederick Jelinek; John D. Lafferty; Robert L. Mercer

In automatic speech recognition, language models can be represented by Probabilistic Context Free Grammars (PCFGs). In this lecture we review some known algorithms which handle PCFGs; in particular an algorithm for the computation of the total probability that a PCFG generates a given sentence (Inside), an algorithm for finding the most probable parse tree (Viterbi), and an algorithm for the estimation of the probabilities of the rewriting rules of a PCFG given a corpus (Inside-Outside). Moreover, we introduce the Left-to-Right Inside algorithm, which computes the probability that successive applications of the grammar rewriting rules (beginning with the sentence start symbol s) produce a word string whose initial substring is a given one.

Collaboration


Dive into the Frederick Jelinek's collaboration.

Researchain Logo
Decentralizing Knowledge