Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Roy Schwartz is active.

Publication


Featured researches published by Roy Schwartz.


conference on computational natural language learning | 2015

Symmetric Pattern Based Word Embeddings for Improved Word Similarity Prediction

Roy Schwartz; Roi Reichart; Ari Rappoport

We present a novel word level vector representation based on symmetric patterns (SPs). For this aim we automatically acquire SPs (e.g., “X and Y”) from a large corpus of plain text, and generate vectors where each coordinate represents the cooccurrence in SPs of the represented word with another word of the vocabulary. Our representation has three advantages over existing alternatives: First, being based on symmetric word relationships, it is highly suitable for word similarity prediction. Particularly, on the SimLex999 word similarity dataset, our model achieves a Spearman’s score of 0.517, compared to 0.462 of the state-of-the-art word2vec model. Interestingly, our model performs exceptionally well on verbs, outperforming stateof-the-art baselines by 20.2‐41.5%. Second, pattern features can be adapted to the needs of a target NLP application. For example, we show that we can easily control whether the embeddings derived from SPs deem antonym pairs (e.g. (big,small)) as similar or dissimilar, an important distinction for tasks such as word classification and sentiment analysis. Finally, we show that a simple combination of the word similarity scores generated by our method and by word2vec results in a superior predictive power over that of each individual model, scoring as high as 0.563 in Spearman’s on SimLex999. This emphasizes the differences between the signals captured by each of the models.


international joint conference on natural language processing | 2015

How Well Do Distributional Models Capture Different Types of Semantic Knowledge

Dana Rubinstein; Effi Levi; Roy Schwartz; Ari Rappoport

In recent years, distributional models (DMs) have shown great success in representing lexical semantics. In this work we show that the extent to which DMs represent semantic knowledge is highly dependent on the type of knowledge. We pose the task of predicting properties of concrete nouns in a supervised setting, and compare between learning taxonomic properties (e.g., animacy) and attributive properties (e.g., size, color). We employ four state-of-the-art DMs as sources of feature representation for this task, and show that they all yield poor results when tested on attributive properties, achieving no more than an average F-score of 0.37 in the binary property prediction task, compared to 0.73 on taxonomic properties. Our results suggest that the distributional hypothesis may not be equally applicable to all types of semantic information.


north american chapter of the association for computational linguistics | 2016

Symmetric Patterns and Coordinations: Fast and Enhanced Representations of Verbs and Adjectives

Roy Schwartz; Roi Reichart; Ari Rappoport

State-of-the-art word embeddings, which are often trained on bag-of-words (BOW) contexts, provide a high quality representation of aspects of the semantics of nouns. However, their quality decreases substantially for the task of verb similarity prediction. In this paper we show that using symmetric pattern contexts (SPs, e.g., “X and Y”) improves word2vec verb similarity performance by up to 15% and is also instrumental in adjective similarity prediction. The unsupervised SP contexts are even superior to a variety of dependency contexts extracted using a supervised dependency parser. Moreover, we observe that SPs and dependency coordination contexts (Coor) capture a similar type of information, and demonstrate that Coor contexts are superior to other dependency contexts including the set of all dependency contexts, although they are still inferior to SPs. Finally, there are substantially fewer SP contexts compared to alternative representations, leading to a massive reduction in training time. On an 8G words corpus and a 32 core machine, the SP model trains in 11 minutes, compared to 5 and 11 hours with BOW and all dependency contexts, respectively.


conference on computational natural language learning | 2017

The Effect of Different Writing Tasks on Linguistic Style: A Case Study of the ROC Story Cloze Task

Roy Schwartz; Maarten Sap; Ioannis Konstas; Leila Zilles; Yejin Choi; Noah A. Smith

A writers style depends not just on personal traits but also on her intent and mental state. In this paper, we show how variants of the same writing task can lead to measurable differences in writing style. We present a case study based on the story cloze task (Mostafazadeh et al., 2016a), where annotators were assigned similar writing tasks with different constraints: (1) writing an entire story, (2) adding a story ending for a given story context, and (3) adding an incoherent ending to a story. We show that a simple linear classifier informed by stylistic features is able to successfully distinguish among the three cases, without even looking at the story context. In addition, combining our stylistic features with language model predictions reaches state of the art performance on the story cloze challenge. Our results demonstrate that different task framings can dramatically affect the way people write.


Proceedings of the 2nd Workshop on Linking Models of Lexical, Sentential and Discourse-level Semantics | 2017

Story Cloze Task: UW NLP System

Roy Schwartz; Maarten Sap; Ioannis Konstas; Leila Zilles; Yejin Choi; Noah A. Smith

This paper describes University of Washington NLP’s submission for the Linking Models of Lexical, Sentential and Discourse-level Semantics (LSDSem 2017) shared task—the Story Cloze Task. Our system is a linear classifier with a variety of features, including both the scores of a neural language model and style features. We report 75.2% accuracy on the task. A further discussion of our results can be found in Schwartz et al. (2017).


conference on computational natural language learning | 2017

Automatic Selection of Context Configurations for Improved Class-Specific Word Representations

Ivan Vulić; Roy Schwartz; Ari Rappoport; Roi Reichart; Anna Korhonen

This paper is concerned with identifying contexts useful for training word representation models for different word classes such as adjectives (A), verbs (V), and nouns (N). We introduce a simple yet effective framework for an automatic selection of class-specific context configurations. We construct a context configuration space based on universal dependency relations between words, and efficiently search this space with an adapted beam search algorithm. In word similarity tasks for each word class, we show that our framework is both effective and efficient. Particularly, it improves the Spearmans rho correlation with human scores on SimLex-999 over the best previously proposed class-specific contexts by 6 (A), 6 (V) and 5 (N) rho points. With our selected context configurations, we train on only 14% (A), 26.2% (V), and 33.6% (N) of all dependency-based contexts, resulting in a reduced training time. Our results generalise: we show that the configurations our algorithm learns for one English training setup outperform previously proposed context types in another training setup for English. Moreover, basing the configuration space on universal dependencies, it is possible to transfer the learned configurations to German and Italian. We also demonstrate improved per-class results over other context types in these two languages.


meeting of the association for computational linguistics | 2011

Neutralizing Linguistically Problematic Annotations in Unsupervised Dependency Parsing Evaluation

Roy Schwartz; Omri Abend; Roi Reichart; Ari Rappoport


international conference on computational linguistics | 2012

Learnability-Based Syntactic Annotation Design

Roy Schwartz; Omri Abend; Ari Rappoport


empirical methods in natural language processing | 2013

Authorship Attribution of Micro-Messages

Roy Schwartz; Oren Tsur; Ari Rappoport; Moshe Koppel


north american chapter of the association for computational linguistics | 2018

ANNOTATION ARTIFACTS IN NATURAL LANGUAGE INFERENCE DATA

Suchin Gururangan; Swabha Swayamdipta; Omer Levy; Roy Schwartz; Samuel R. Bowman; Noah A. Smith

Collaboration


Dive into the Roy Schwartz's collaboration.

Top Co-Authors

Avatar

Ari Rappoport

Hebrew University of Jerusalem

View shared research outputs
Top Co-Authors

Avatar

Roi Reichart

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Noah A. Smith

University of Washington

View shared research outputs
Top Co-Authors

Avatar

Omri Abend

Hebrew University of Jerusalem

View shared research outputs
Top Co-Authors

Avatar

Yejin Choi

University of Washington

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Maarten Sap

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Sam Thomson

Carnegie Mellon University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dana Rubinstein

Hebrew University of Jerusalem

View shared research outputs
Researchain Logo
Decentralizing Knowledge