Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Wojciech Zaremba is active.

Publication


Featured researches published by Wojciech Zaremba.


international joint conference on natural language processing | 2015

Addressing the Rare Word Problem in Neural Machine Translation

Thang Luong; Ilya Sutskever; Quoc V. Le; Oriol Vinyals; Wojciech Zaremba

Neural Machine Translation (NMT) is a new approach to machine translation that has shown promising results that are comparable to traditional approaches. A significant weakness in conventional NMT systems is their inability to correctly translate very rare words: end-to-end NMTs tend to have relatively small vocabularies with a single unk symbol that represents every possible out-of-vocabulary (OOV) word. In this paper, we propose and implement an effective technique to address this problem. We train an NMT system on data that is augmented by the output of a word alignment algorithm, allowing the NMT system to emit, for each OOV word in the target sentence, the position of its corresponding word in the source sentence. This information is later utilized in a post-processing step that translates every OOV word using a dictionary. Our experiments on the WMT’14 English to French translation task show that this method provides a substantial improvement of up to 2.8 BLEU points over an equivalent NMT system that does not use this technique. With 37.5 BLEU points, our NMT system is the first to surpass the best result achieved on a WMT’14 contest task.


workshop on applications of computer vision | 2016

Discriminative training of CRF models with probably submodular constraints

Wojciech Zaremba; Matthew B. Blaschko

Problems of segmentation, denoising, registration and 3d reconstruction are often addressed with the graph cut algorithm. However, solving an unconstrained graph cut problem is NP-hard. For tractable optimization, pairwise potentials have to fulfill the submodularity inequality. In our learning paradigm, pairwise potentials are created as the dot product of a learned vector w with positive feature vectors. In order to constrain such a model to remain tractable, previous approaches have enforced the weight vector to be positive for pairwise potentials in which the labels differ, and set pairwise potentials to zero in the case that the label remains the same. Such constraints are sufficient to guarantee that the resulting pairwise potentials satisfy the submodularity inequality. However, we show that such an approach unnecessarily restricts the capacity of the learned models. Instead, we approach the problem of learning with sub-modularity constraints from a probabilistic setting. Prediction errors may be the result of: learning error, model error, or inference error. Guaranteeing submodularity for all possible inputs, no matter how improbable, reduces inference error to effectively zero, but increases model error. In contrast, we relax the requirement of guaranteed submodularity to solutions that are submodular with high probability. We show that the conceptually simple strategy of enforcing submodularity on the training examples guarantees with low sample complexity that test images will also yield sub-modular pairwise potentials. Results are presented showing substantial improvement from the resulting increased model capacity.


international conference on learning representations | 2014

Intriguing properties of neural networks

Christian Szegedy; Wojciech Zaremba; Ilya Sutskever; Joan Bruna; Dumitru Erhan; Ian J. Goodfellow; Rob Fergus


arXiv: Neural and Evolutionary Computing | 2014

Recurrent Neural Network Regularization

Wojciech Zaremba; Ilya Sutskever; Oriol Vinyals


international conference on machine learning | 2015

An Empirical Exploration of Recurrent Network Architectures

Rafal Jozefowicz; Wojciech Zaremba; Ilya Sutskever


neural information processing systems | 2014

Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation

Emily L. Denton; Wojciech Zaremba; Joan Bruna; Yann LeCun; Rob Fergus


international conference on learning representations | 2014

Spectral Networks and Locally Connected Networks on Graphs

Joan Bruna; Wojciech Zaremba; Arthur Szlam; Yann LeCun


arXiv: Neural and Evolutionary Computing | 2015

Learning to Execute

Wojciech Zaremba; Ilya Sutskever


Archive | 2015

Reinforcement Learning Neural Turing Machines

Wojciech Zaremba; Ilya Sutskever


international conference on machine learning | 2016

Learning simple algorithms from examples

Wojciech Zaremba; Tomas Mikolov; Armand Joulin; Rob Fergus

Collaboration


Dive into the Wojciech Zaremba's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Arthur Szlam

City College of New York

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge