Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Markos Mylonakis is active.

Publication


Featured researches published by Markos Mylonakis.


empirical methods in natural language processing | 2008

Phrase Translation Probabilities with ITG Priors and Smoothing as Learning Objective

Markos Mylonakis; Khalil Sima'an

The conditional phrase translation probabilities constitute the principal components of phrase-based machine translation systems. These probabilities are estimated using a heuristic method that does not seem to optimize any reasonable objective function of the word-aligned, parallel training corpus. Earlier efforts on devising a better understood estimator either do not scale to reasonably sized training data, or lead to deteriorating performance. In this paper we explore a new approach based on three ingredients (1) A generative model with a prior over latent segmentations derived from Inversion Transduction Grammar (ITG), (2) A phrase table containing all phrase pairs without length limit, and (3) Smoothing as learning objective using a novel Maximum-A-Posteriori version of Deleted Estimation working with Expectation-Maximization. Where others conclude that latent segmentations lead to overfitting and deteriorating performance, we show here that these three ingredients give performance equivalent to the heuristic method on reasonably sized training data.


international conference on machine learning | 2007

Unsupervised estimation for noisy-channel models

Markos Mylonakis; Khalil Sima'an; Rebecca Hwa

Shannons Noisy-Channel model, which describes how a corrupted message might be reconstructed, has been the corner stone for much work in statistical language and speech processing. The model factors into two components: a language model to characterize the original message and a channel model to describe the channels corruptive process. The standard approach for estimating the parameters of the channel model is unsupervised Maximum-Likelihood of the observation data, usually approximated using the Expectation-Maximization (EM) algorithm. In this paper we show that it is better to maximize the joint likelihood of the data at both ends of the noisy-channel. We derive a corresponding bi-directional EM algorithm and show that it gives better performance than standard EM on two tasks: (1) translation using a probabilistic lexicon and (2) adaptation of a part-of-speech tagger between related languages.


spoken language technology workshop | 2008

Better statistical estimation can benefit all phrases in phrase-based statistical machine translation

Khalil Sima'an; Markos Mylonakis

The heuristic estimates of conditional phrase translation probabilities are based on frequency counts in a word-aligned parallel corpus. Earlier attempts at more principled estimation using Expectation-Maximization (EM) under perform this heuristic. This paper shows that a recently introduced novel estimator based on smoothing might provide a good alternative. When all phrase pairs are estimated (no length cut-off), this estimator slightly outperforms the heuristic estimator.


Journal of Logic and Computation | 2014

Learning structural dependencies of words in the Zipfian Tail

Tejaswini Deoskar; Markos Mylonakis; Khalil Sima'an

Using semi-supervised EM, we learn finegrained but sparse lexical parameters of a generative parsing model (a PCFG) initially estimated over the Penn Treebank. Our lexical parameters employ supertags, which encode complex structural information at the pre-terminal level, and are particularly sparse in labeled data -- our goal is to learn these for words that are unseen or rare in the labeled data. In order to guide estimation from unlabeled data, we incorporate both structural and lexical priors from the labeled data. We get a large error reduction in parsing ambiguous structures associated with unseen verbs, the most important case of learning lexico-structural dependencies. We also obtain a statistically significant improvement in labeled bracketing score of the treebank PCFG, the first successful improvement via semi-supervised EM of a generative structured model already trained over large labeled data.


meeting of the association for computational linguistics | 2011

Learning Hierarchical Translation Structure with Linguistic Annotations

Markos Mylonakis; Khalil Sima'an


conference on computational natural language learning | 2010

Learning Probabilistic Synchronous CFGs for Phrase-Based Translation

Markos Mylonakis; Khalil Sima'an


Archive | 2012

Learning the latent structure of translation

Markos Mylonakis


international workshop/conference on parsing technologies | 2011

Learning Structural Dependencies of Words in the Zipfian Tail

Tejaswini Deoskar; Markos Mylonakis; Khalil Sima'an


Convergence | 2008

BETTER STATISTICAL ESTIMATION CAN BENEFIT ALL PHRASES IN PHRASE-BASED STATISTICAL MACHINE TRANSLATION

Khalil Sima'an; Markos Mylonakis


Archive | 2007

Translation Lexicon Estimates from Non-Parallel Corpora Pairs

Markos Mylonakis; Khalil Sima'an; Dastani, M., de Jong, E.

Collaboration


Dive into the Markos Mylonakis's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Rebecca Hwa

University of Pittsburgh

View shared research outputs
Researchain Logo
Decentralizing Knowledge