Nal Kalchbrenner
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Nal Kalchbrenner.
Nature | 2016
David Silver; Aja Huang; Chris J. Maddison; Arthur Guez; Laurent Sifre; George van den Driessche; Julian Schrittwieser; Ioannis Antonoglou; Veda Panneershelvam; Marc Lanctot; Sander Dieleman; Dominik Grewe; John Nham; Nal Kalchbrenner; Ilya Sutskever; Timothy P. Lillicrap; Madeleine Leach; Koray Kavukcuoglu; Thore Graepel; Demis Hassabis
The game of Go has long been viewed as the most challenging of classic games for artificial intelligence owing to its enormous search space and the difficulty of evaluating board positions and moves. Here we introduce a new approach to computer Go that uses ‘value networks’ to evaluate board positions and ‘policy networks’ to select moves. These deep neural networks are trained by a novel combination of supervised learning from human expert games, and reinforcement learning from games of self-play. Without any lookahead search, the neural networks play Go at the level of state-of-the-art Monte Carlo tree search programs that simulate thousands of random games of self-play. We also introduce a new search algorithm that combines Monte Carlo simulation with value and policy networks. Using this search algorithm, our program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0. This is the first time that a computer program has defeated a human professional player in the full-sized game of Go, a feat previously thought to be at least a decade away.
meeting of the association for computational linguistics | 2014
Nal Kalchbrenner; Edward Grefenstette; Phil Blunsom
The ability to accurately represent sentences is central to language understanding. We describe a convolutional architecture dubbed the Dynamic Convolutional Neural Network (DCNN) that we adopt for the semantic modelling of sentences. The network uses Dynamic k-Max Pooling, a global pooling operation over linear sequences. The network handles input sentences of varying length and induces a feature graph over the sentence that is capable of explicitly capturing short and long-range relations. The network does not rely on a parse tree and is easily applicable to any language. We test the DCNN in four experiments: small scale binary and multi-class sentiment prediction, six-way question classification and Twitter sentiment prediction by distant supervision. The network achieves excellent performance in the first three tasks and a greater than 25% error reduction in the last task with respect to the strongest baseline.
meeting of the association for computational linguistics | 2014
Dimitri Kartsaklis; Nal Kalchbrenner; Mehrnoosh Sadrzadeh
This paper provides a method for improving tensor-based compositional distributional models of meaning by the addition of an explicit disambiguation step prior to composition. In contrast with previous research where this hypothesis has been successfully tested against relatively simple compositional models, in our work we use a robust model trained with linear regression. The results we get in two experiments show the superiority of the prior disambiguation method and suggest that the effectiveness of this approach is modelindependent.
empirical methods in natural language processing | 2013
Nal Kalchbrenner; Phil Blunsom
arXiv: Sound | 2016
Aäron van den Oord; Sander Dieleman; Heiga Zen; Karen Simonyan; Oriol Vinyals; Alex Graves; Nal Kalchbrenner; Andrew W. Senior; Koray Kavukcuoglu
international conference on machine learning | 2016
Aäron van den Oord; Nal Kalchbrenner; Koray Kavukcuoglu
neural information processing systems | 2016
Aäron van den Oord; Nal Kalchbrenner; Lasse Espeholt; Koray Kavukcuoglu; Oriol Vinyals; Alex Graves
international conference on learning representations | 2016
Nal Kalchbrenner; Ivo Danihelka; Alex Graves
meeting of the association for computational linguistics | 2013
Nal Kalchbrenner; Phil Blunsom
arXiv: Computation and Language | 2016
Nal Kalchbrenner; Lasse Espeholt; Karen Simonyan; Aäron van den Oord; Alex Graves; Koray Kavukcuoglu