Alexis Conneau
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Alexis Conneau.
empirical methods in natural language processing | 2017
Alexis Conneau; Douwe Kiela; Holger Schwenk; Loïc Barrault; Antoine Bordes
Many modern NLP systems rely on word embeddings, previously trained in an unsupervised manner on large corpora, as base features. Efforts to obtain embeddings for larger chunks of text, such as sentences, have however not been so successful. Several attempts at learning unsupervised representations of sentences have not reached satisfactory enough performance to be widely adopted. In this paper, we show how universal sentence representations trained using the supervised data of the Stanford Natural Language Inference datasets can consistently outperform unsupervised methods like SkipThought vectors on a wide range of transfer tasks. Much like how computer vision uses ImageNet to obtain features, which can then be transferred to other tasks, our work tends to indicate the suitability of natural language inference for transfer learning to other NLP tasks. Our encoder is publicly available.
conference on recommender systems | 2016
Flavian Vasile; Elena Smirnova; Alexis Conneau
We propose Meta-Prod2vec, a novel method to compute item similarities for recommendation that leverages existing item metadata. Such scenarios are frequently encountered in applications such as content recommendation, ad targeting and web search. Our method leverages past user interactions with items and their attributes to compute low-dimensional embeddings of items. Specifically, the item metadata is injected into the model as side information to regularize the item embeddings. We show that the new item representations lead to better performance on recommendation tasks on an open music dataset.
north american chapter of the association for computational linguistics | 2018
Douwe Kiela; Alexis Conneau; Allan Jabri; Maximilian Nickel
We introduce a variety of models, trained on a supervised image captioning corpus to predict the image features for a given caption, to perform sentence representation grounding. We train a grounded sentence encoder that achieves good performance on COCO caption and image retrieval and subsequently show that this encoder can successfully be transferred to various NLP tasks, with improved performance over text-only models. Lastly, we analyze the contribution of grounding, and show that word embeddings learned by this system outperform non-grounded ones.
arXiv: Computation and Language | 2016
Alexis Conneau; Holger Schwenk; Loïc Barrault; Yann LeCun
conference of the european chapter of the association for computational linguistics | 2017
Alexis Conneau; Holger Schwenk; Loïc Barrault; Yann LeCun
international conference on learning representations | 2018
Guillaume Lample; Alexis Conneau; Ludovic Denoyer; Marc'Aurelio Ranzato
international conference on learning representations | 2018
Guillaume Lample; Alexis Conneau; Marc'Aurelio Ranzato; Ludovic Denoyer; Hervé Jégou
meeting of the association for computational linguistics | 2018
Alexis Conneau; Germán Kruszewski; Guillaume Lample; Loïc Barrault; Marco Baroni
language resources and evaluation | 2018
Alexis Conneau; Douwe Kiela
empirical methods in natural language processing | 2018
Guillaume Lample; Myle Ott; Alexis Conneau; Ludovic Denoyer; Marc'Aurelio Ranzato