Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yulia Tsvetkov is active.

Publication


Featured researches published by Yulia Tsvetkov.


empirical methods in natural language processing | 2015

Not All Contexts Are Created Equal: Better Word Representations with Variable Attention

Wang Ling; Yulia Tsvetkov; Silvio Amir; Ramon Fermandez; Chris Dyer; Alan W. Black; Isabel Trancoso; Chu-Cheng Lin

We introduce an extension to the bag-ofwords model for learning words representations that take into account both syntactic and semantic properties within language. This is done by employing an attention model that finds within the contextual words, the words that are relevant for each prediction. The general intuition of our model is that some words are only relevant for predicting local context (e.g. function words), while other words are more suited for determining global context, such as the topic of the document. Experiments performed on both semantically and syntactically oriented tasks show gains using our model over the existing bag of words model. Furthermore, compared to other more sophisticated models, our model scales better as we increase the size of the context of the model.


empirical methods in natural language processing | 2015

Evaluation of Word Vector Representations by Subspace Alignment

Yulia Tsvetkov; Manaal Faruqui; Wang Ling; Guillaume Lample; Chris Dyer

Unsupervisedly learned word vectors have proven to provide exceptionally effective features in many NLP tasks. Most common intrinsic evaluations of vector quality measure correlation with similarity judgments. However, these often correlate poorly with how well the learned representations perform as features in downstream evaluation tasks. We present QVEC—a computationally inexpensive intrinsic evaluation measure of the quality of word embeddings based on alignment to a matrix of features extracted from manually crafted lexical resources—that obtains strong correlation with performance of the vectors in a battery of downstream semantic evaluation tasks.1


meeting of the association for computational linguistics | 2014

Metaphor Detection with Cross-Lingual Model Transfer

Yulia Tsvetkov; Leonid Boytsov; Anatole Gershman; Eric Nyberg; Chris Dyer

We show that it is possible to reliably discriminate whether a syntactic construction is meant literally or metaphorically using lexical semantic features of the words that participate in the construction. Our model is constructed using English resources, and we obtain state-of-the-art performance relative to previous work in this language. Using a model transfer approach by pivoting through a bilingual dictionary, we show our model can identify metaphoric expressions in other languages. We provide results on three new test sets in Spanish, Farsi, and Russian. The results support the hypothesis that metaphors are conceptual, rather than lexical, in nature.


international joint conference on natural language processing | 2015

Sparse Overcomplete Word Vector Representations

Manaal Faruqui; Yulia Tsvetkov; Dani Yogatama; Chris Dyer; Noah A. Smith

Current distributed representations of words show little resemblance to theories of lexical semantics. The former are dense and uninterpretable, the latter largely based on familiar, discrete classes (e.g., supersenses) and relations (e.g., synonymy and hypernymy). We propose methods that transform word vectors into sparse (and optionally binary) vectors. The resulting representations are more similar to the interpretable features typically used in NLP, though they are discovered automatically from raw corpora. Because the vectors are highly sparse, they are computationally easy to work with. Most importantly, we find that they outperform the original vectors on benchmark tasks.


workshop on evaluating vector space representations for nlp | 2016

Problems With Evaluation of Word Embeddings Using Word Similarity Tasks

Manaal Faruqui; Yulia Tsvetkov; Pushpendre Rastogi; Chris Dyer

Lacking standardized extrinsic evaluation methods for vector representations of words, the NLP community has relied heavily on word similarity tasks as a proxy for intrinsic evaluation of word vectors. Word similarity evaluation, which correlates the distance between vectors and human judgments of semantic similarity is attractive, because it is computationally inexpensive and fast. In this paper we present several problems associated with the evaluation of word vectors on word similarity datasets, and summarize existing solutions. Our study suggests that the use of word similarity tasks for evaluation of word vectors is not sustainable and calls for further research on evaluation methods.


empirical methods in natural language processing | 2011

Identification of Multi-word Expressions by Combining Multiple Linguistic Information Sources

Yulia Tsvetkov; Shuly Wintner

We propose a framework for using multiple sources of linguistic information in the task of identifying multiword expressions in natural language texts. We define various linguistically motivated classification features and introduce novel ways for computing them. We then manually define interrelationships among the features, and express them in a Bayesian network. The result is a powerful classifier that can identify multiword expressions of various types and multiple syntactic constructions in text corpora. Our methodology is unsupervised and language-independent; it requires relatively few language resources and is thus suitable for a large number of languages. We report results on English, French, and Hebrew, and demonstrate a significant improvement in identification accuracy, compared with less sophisticated baselines.


Natural Language Engineering | 2011

Extraction of multi-word expressions from small parallel corpora

Yulia Tsvetkov; Shuly Wintner

We present a general, novel methodology for extracting multi-word expressions (MWEs) of various types, along with their translations, from small, word-aligned parallel corpora. Unlike existing approaches, we focus on misalignments; these typically indicate expressions in the source language that are translated to the target in a non-compositional way. We introduce a simple algorithm that proposes MWE candidates based on such misalignments, relying on 1:1 alignments as anchors that delimit the search space. We use a large monolingual corpus to rank and filter these candidates. Evaluation of the quality of the extraction algorithm reveals significant improvements over naive alignment-based methods. The extracted MWEs, with their translations, are used in the training of a statistical machine translation system, showing a small but significant improvement in its performance.


north american chapter of the association for computational linguistics | 2016

Polyglot Neural Language Models: A Case Study in Cross-Lingual Phonetic Representation Learning

Yulia Tsvetkov; Sunayana Sitaram; Manaal Faruqui; Guillaume Lample; Patrick Littell; David R. Mortensen; Alan W. Black; Lori S. Levin; Chris Dyer

We introduce polyglot language models, recurrent neural network models trained to predict symbol sequences in many different languages using shared representations of symbols and conditioning on typological information about the language to be predicted. We apply these to the problem of modeling phone sequences---a domain in which universal symbol inventories and cross-linguistically shared feature representations are a natural fit. Intrinsic evaluation on held-out perplexity, qualitative analysis of the learned representations, and extrinsic evaluation in two downstream applications that make use of phonetic features show (i) that polyglot models better generalize to held-out data than comparable monolingual models and (ii) that polyglot phonetic feature representations are of higher quality than those learned monolingually.


north american chapter of the association for computational linguistics | 2016

Morphological Inflection Generation Using Character Sequence to Sequence Learning

Manaal Faruqui; Yulia Tsvetkov; Graham Neubig; Chris Dyer

Morphological inflection generation is the task of generating the inflected form of a given lemma corresponding to a particular linguistic transformation. We model the problem of inflection generation as a character sequence to sequence learning problem and present a variant of the neural encoder-decoder model for solving it. Our model is language independent and can be trained in both supervised and semi-supervised settings. We evaluate our system on seven datasets of morphologically rich languages and achieve either better or comparable results to existing state-of-the-art models of inflection generation.


north american chapter of the association for computational linguistics | 2015

Constraint-Based Models of Lexical Borrowing.

Yulia Tsvetkov; Waleed Ammar; Chris Dyer

Linguistic borrowing is the phenomenon of transferring linguistic constructions (lexical, phonological, morphological, and syntactic) from a “donor” language to a “recipient” language as a result of contacts between communities speaking different languages. Borrowed words are found in all languages, and—in contrast to cognate relationships—borrowing relationships may exist across unrelated languages (for example, about 40% of Swahili’s vocabulary is borrowed from Arabic). In this paper, we develop a model of morpho-phonological transformations across languages with features based on universal constraints from Optimality Theory (OT). Compared to several standard— but linguistically naive—baselines, our OTinspired model obtains good performance with only a few dozen training examples, making this a cost-effective strategy for sharing lexical information across languages.

Collaboration


Dive into the Yulia Tsvetkov's collaboration.

Top Co-Authors

Avatar

Chris Dyer

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Manaal Faruqui

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lori S. Levin

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Nathan Schneider

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Alan W. Black

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Chu-Cheng Lin

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Guillaume Lample

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar

Waleed Ammar

Carnegie Mellon University

View shared research outputs
Researchain Logo
Decentralizing Knowledge