Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Rémi Lebret is active.

Publication


Featured researches published by Rémi Lebret.


conference of the european chapter of the association for computational linguistics | 2014

Word Embeddings through Hellinger PCA

Rémi Lebret; Ronan Collobert

Word embeddings resulting from neural language models have been shown to be a great asset for a large variety of NLP tasks. However, such architecture might be difficult and time-consuming to train. Instead, we propose to drastically simplify the word embeddings computation through a Hellinger PCA of the word co- occurence matrix. We compare those new word embeddings with some well-known embeddings on named entity recognition and movie review tasks and show that we can reach similar or even better performance. Although deep learning is not really necessary for generating good word embeddings, we show that it can provide an easy way to adapt embeddings to specific tasks.


empirical methods in natural language processing | 2016

Neural Text Generation from Structured Data with Application to the Biography Domain.

Rémi Lebret; Michael Auli

This paper introduces a neural model for concept-to-text generation that scales to large, rich domains. We experiment with a new dataset of biographies from Wikipedia that is an order of magnitude larger than existing resources with over 700k samples. The dataset is also vastly more diverse with a 400k vocabulary, compared to a few hundred words for Weathergov or Robocup. Our model builds upon recent work on conditional neural language model for text generation. To deal with the large vocabulary, we extend these models to mix a fixed vocabulary with copy actions that transfer sample-specific words from the input database to the generated output sentence. Our neural model significantly out-performs a classical Kneser-Ney language model adapted to this task by nearly 15 BLEU.


conference on intelligent text processing and computational linguistics | 2015

Rehabilitation of Count-based Models for Word Vector Representations

Rémi Lebret; Ronan Collobert

Recent works on word representations mostly rely on predictive models. Distributed word representations (aka word embeddings) are trained to optimally predict the contexts in which the corresponding words tend to appear. Such models have succeeded in capturing word similarities as well as semantic and syntactic regularities. Instead, we aim at reviving interest in a model based on counts. We present a systematic study of the use of the Hellinger distance to extract semantic representations from the word co-occurrence statistics of large text corpora. We show that this distance gives good performance on word similarity and analogy tasks, with a proper type and size of context, and a dimensionality reduction based on a stochastic low-rank approximation. Besides being both simple and intuitive, this method also provides an encoding function which can be used to infer unseen words or phrases. This becomes a clear advantage compared to predictive models which must train these new words.


conference on information and knowledge management | 2017

Taxonomy Induction Using Hypernym Subsequences

Amit Gupta; Rémi Lebret; Hamza Harkous; Karl Aberer

We propose a novel, semi-supervised approach towards domain taxonomy induction from an input vocabulary of seed terms. Unlike all previous approaches, which typically extract direct hypernym edges for terms, our approach utilizes a novel probabilistic framework to extract hypernym subsequences. Taxonomy induction from extracted subsequences is cast as an instance of the minimum-cost flow problem on a carefully designed directed graph. Through experiments, we demonstrate that our approach outperforms state-of-the-art taxonomy induction approaches across four languages. Importantly, we also show that our approach is robust to the presence of noise in the input vocabulary. To the best of our knowledge, this robustness has not been empirically proven in any previous approach.


international conference on machine learning | 2015

Phrase-based Image Captioning

Rémi Lebret; Pedro H. O. Pinheiro; Ronan Collobert


Archive | 2013

Is Deep Learning Really Necessary for Word Embeddings

Rémi Lebret; Joël Legrand; Ronan Collobert


arXiv: Computation and Language | 2014

Simple Image Description Generator via a Linear Phrase-Based Approach.

Rémi Lebret; Pedro H. O. Pinheiro; Ronan Collobert


arXiv: Computation and Language | 2016

Generating Text from Structured Data with Application to the Biography Domain.

Rémi Lebret; Michael Auli


international conference on learning representations | 2015

N-gram-Based Low-Dimensional Representation for Document Classification

Rémi Lebret; Ronan Collobert


usenix security symposium | 2018

Polisis: Automated Analysis and Presentation of Privacy Policies Using Deep Learning.

Hamza Harkous; Kassem Fawaz; Rémi Lebret; Florian Schaub; Kang G. Shin; Karl Aberer

Collaboration


Dive into the Rémi Lebret's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Karl Aberer

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Hamza Harkous

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Amit Gupta

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Pedro H. O. Pinheiro

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Michael Auli

University of Edinburgh

View shared research outputs
Top Co-Authors

Avatar

Chi Thang Duong

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar

Joël Legrand

Idiap Research Institute

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge