Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Armand Joulin is active.

Publication


Featured researches published by Armand Joulin.


european conference on computer vision | 2016

Learning Visual Features from Large Weakly Supervised Data

Armand Joulin; Laurens van der Maaten; Allan Jabri; Nicolas Vasilache

Convolutional networks trained on large supervised datasets produce visual features which form the basis for the state-of-the-art in many computer-vision problems. Further improvements of these visual features will likely require even larger manually labeled data sets, which severely limits the pace at which progress can be made. In this paper, we explore the potential of leveraging massive, weakly-labeled image collections for learning good visual features. We train convolutional networks on a dataset of 100 million Flickr photos and comments, and show that these networks produce features that perform well in a range of vision problems. We also show that the networks appropriately capture word similarity and learn correspondences between different languages.


european conference on computer vision | 2016

Revisiting Visual Question Answering Baselines

Allan Jabri; Armand Joulin; Laurens van der Maaten

Visual question answering (VQA) is an interesting learning setting for evaluating the abilities and shortcomings of current systems for image understanding. Many of the recently proposed VQA systems include attention or memory mechanisms designed to perform “reasoning”. Furthermore, for the task of multiple-choice VQA, nearly all of these systems train a multi-class classifier on image and question features to predict an answer. This paper questions the value of these common practices and develops a simple alternative model based on binary classification. Instead of treating answers as competing choices, our model receives the answer as input and predicts whether or not an image-question-answer triplet is correct. We evaluate our model on the Visual7W Telling and the VQA Real Multiple Choice tasks, and find that even simple versions of our model perform competitively. Our best model achieves state-of-the-art performance of \(65.8\,\%\) accuracy on the Visual7W Telling task and compares surprisingly well with the most complex systems proposed for the VQA Real Multiple Choice task. Additionally, we explore variants of the model and study the transferability of the model between both datasets. We also present an error analysis of our best model, the results of which suggest that a key problem of current VQA systems lies in the lack of visual grounding and localization of concepts that occur in the questions and answers.


conference on intelligent text processing and computational linguistics | 2016

A Roadmap Towards Machine Intelligence

Tomas Mikolov; Armand Joulin; Marco Baroni

The development of intelligent machines is one of the biggest unsolved challenges in computer science. In this paper, we propose some fundamental properties these machines should have, focusing in particular on communication and learning. We discuss a simple environment that could be used to incrementally teach a machine the basics of natural-language-based communication, as a prerequisite to more complex interaction with human users. We also present some conjectures on the sort of algorithms the machine should support in order to profitably learn from the environment.


international conference on learning representations | 2016

Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks

Jason Weston; Antoine Bordes; Sumit Chopra; Alexander M. Rush; Bart van Merriënboer; Armand Joulin; Tomas Mikolov


Transactions of the Association for Computational Linguistics | 2017

Enriching Word Vectors with Subword Information

Piotr Bojanowski; Edouard Grave; Armand Joulin; Tomas Mikolov


neural information processing systems | 2015

Inferring algorithmic patterns with stack-augmented recurrent nets

Armand Joulin; Tomas Mikolov


arXiv: Neural and Evolutionary Computing | 2014

Learning Longer Memory in Recurrent Neural Networks.

Tomas Mikolov; Armand Joulin; Sumit Chopra; Michael Mathieu; Marc'Aurelio Ranzato


international conference on machine learning | 2016

Learning simple algorithms from examples

Wojciech Zaremba; Tomas Mikolov; Armand Joulin; Rob Fergus


international conference on learning representations | 2017

Improving Neural Language Models with a Continuous Cache

Edouard Grave; Armand Joulin; Nicolas Usunier


international conference on machine learning | 2017

Efficient Softmax Approximation for GPUs

Edouard Grave; Armand Joulin; Moustapha Cisse; David Grangier; Hervé Jégou

Collaboration


Dive into the Armand Joulin's collaboration.

Top Co-Authors

Avatar

Edouard Grave

University of California

View shared research outputs
Top Co-Authors

Avatar

Piotr Bojanowski

French Institute for Research in Computer Science and Automation

View shared research outputs
Top Co-Authors

Avatar

Laurens van der Maaten

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge