Wojciech Zaremba
New York University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Wojciech Zaremba.
international joint conference on natural language processing | 2015
Thang Luong; Ilya Sutskever; Quoc V. Le; Oriol Vinyals; Wojciech Zaremba
Neural Machine Translation (NMT) is a new approach to machine translation that has shown promising results that are comparable to traditional approaches. A significant weakness in conventional NMT systems is their inability to correctly translate very rare words: end-to-end NMTs tend to have relatively small vocabularies with a single unk symbol that represents every possible out-of-vocabulary (OOV) word. In this paper, we propose and implement an effective technique to address this problem. We train an NMT system on data that is augmented by the output of a word alignment algorithm, allowing the NMT system to emit, for each OOV word in the target sentence, the position of its corresponding word in the source sentence. This information is later utilized in a post-processing step that translates every OOV word using a dictionary. Our experiments on the WMT’14 English to French translation task show that this method provides a substantial improvement of up to 2.8 BLEU points over an equivalent NMT system that does not use this technique. With 37.5 BLEU points, our NMT system is the first to surpass the best result achieved on a WMT’14 contest task.
workshop on applications of computer vision | 2016
Wojciech Zaremba; Matthew B. Blaschko
Problems of segmentation, denoising, registration and 3d reconstruction are often addressed with the graph cut algorithm. However, solving an unconstrained graph cut problem is NP-hard. For tractable optimization, pairwise potentials have to fulfill the submodularity inequality. In our learning paradigm, pairwise potentials are created as the dot product of a learned vector w with positive feature vectors. In order to constrain such a model to remain tractable, previous approaches have enforced the weight vector to be positive for pairwise potentials in which the labels differ, and set pairwise potentials to zero in the case that the label remains the same. Such constraints are sufficient to guarantee that the resulting pairwise potentials satisfy the submodularity inequality. However, we show that such an approach unnecessarily restricts the capacity of the learned models. Instead, we approach the problem of learning with sub-modularity constraints from a probabilistic setting. Prediction errors may be the result of: learning error, model error, or inference error. Guaranteeing submodularity for all possible inputs, no matter how improbable, reduces inference error to effectively zero, but increases model error. In contrast, we relax the requirement of guaranteed submodularity to solutions that are submodular with high probability. We show that the conceptually simple strategy of enforcing submodularity on the training examples guarantees with low sample complexity that test images will also yield sub-modular pairwise potentials. Results are presented showing substantial improvement from the resulting increased model capacity.
international conference on learning representations | 2014
Christian Szegedy; Wojciech Zaremba; Ilya Sutskever; Joan Bruna; Dumitru Erhan; Ian J. Goodfellow; Rob Fergus
arXiv: Neural and Evolutionary Computing | 2014
Wojciech Zaremba; Ilya Sutskever; Oriol Vinyals
international conference on machine learning | 2015
Rafal Jozefowicz; Wojciech Zaremba; Ilya Sutskever
neural information processing systems | 2014
Emily L. Denton; Wojciech Zaremba; Joan Bruna; Yann LeCun; Rob Fergus
international conference on learning representations | 2014
Joan Bruna; Wojciech Zaremba; Arthur Szlam; Yann LeCun
arXiv: Neural and Evolutionary Computing | 2015
Wojciech Zaremba; Ilya Sutskever
Archive | 2015
Wojciech Zaremba; Ilya Sutskever
international conference on machine learning | 2016
Wojciech Zaremba; Tomas Mikolov; Armand Joulin; Rob Fergus