Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Kurt Driessens is active.

Publication


Featured researches published by Kurt Driessens.


inductive logic programming | 2006

Graph kernels and Gaussian processes for relational reinforcement learning

Kurt Driessens; Jan Ramon; Thomas Gärtner

RRL is a relational reinforcement learning system based on Q-learning in relational state-action spaces. It aims to enable agents to learn how to act in an environment that has no natural representation as a tuple of constants. For relational reinforcement learning, the learning algorithm used to approximate the mapping between state-action pairs and their so called Q(uality)-value has to be very reliable, and it has to be able to handle the relational representation of state-action pairs. In this paper we investigate the use of Gaussian processes to approximate the Q-values of state-action pairs. In order to employ Gaussian processes in a relational setting we propose graph kernels as a covariance function between state-action pairs. The standard prediction mechanism for Gaussian processes requires a matrix inversion which can become unstable when the kernel matrix has low rank. These instabilities can be avoided by employing QR-factorization. This leads to better and more stable performance of the algorithm and a more efficient incremental update mechanism. Experiments conducted in the blocks world and with the Tetris game show that Gaussian processes with graph kernels can compete with, and often improve on, regression trees and instance based regression as a generalization algorithm for RRL.


european conference on machine learning | 2001

Speeding Up Relational Reinforcement Learning through the Use of an Incremental First Order Decision Tree Learner

Kurt Driessens; Jan Ramon; Hendrik Blockeel

Relational reinforcement learning (RRL) is a learning technique that combines standard reinforcement learning with inductive logic programming to enable the learning system to exploit structural knowledge about the application domain. This paper discusses an improvement of the original RRL. We introduce a fully incremental first order decision tree learning algorithm TG and integrate this algorithm in the RRL system to form RRL-TG. We demonstrate the performance gain on similar experiments to those that were used to demonstrate the behaviour of the original RRL system.


Machine Learning | 2004

Integrating Guidance into Relational Reinforcement Learning

Kurt Driessens; Sašo Džeroski

Reinforcement learning, and Q-learning in particular, encounter two major problems when dealing with large state spaces. First, learning the Q-function in tabular form may be infeasible because of the excessive amount of memory needed to store the table, and because the Q-function only converges after each state has been visited multiple times. Second, rewards in the state space may be so sparse that with random exploration they will only be discovered extremely slowly. The first problem is often solved by learning a generalization of the encountered examples (e.g., using a neural net or decision tree). Relational reinforcement learning (RRL) is such an approach; it makes Q-learning feasible in structural domains by incorporating a relational learner into Q-learning. The problem of sparse rewards has not been addressed for RRL. This paper presents a solution based on the use of “reasonable policies” to provide guidance. Different types of policies and different strategies to supply guidance through these policies are discussed and evaluated experimentally in several relational domains to show the merits of the approach.


international conference on machine learning | 2008

Non-parametric policy gradients: a unified treatment of propositional and relational domains

Kristian Kersting; Kurt Driessens

Policy gradient approaches are a powerful instrument for learning how to interact with the environment. Existing approaches have focused on propositional and continuous domains only. Without extensive feature engineering, it is difficult - if not impossible - to apply them within structured domains, in which e.g. there is a varying number of objects and relations among them. In this paper, we describe a non-parametric policy gradient approach - called NPPG - that overcomes this limitation. The key idea is to apply Friedmanns gradient boosting: policies are represented as a weighted sum of regression models grown in an stage-wise optimization. Employing off-the-shelf regression learners, NPPG can deal with propositional, continuous, and relational domains in a unified way. Our experimental results show that it can even improve on established results.


european conference on machine learning | 2007

Transfer Learning in Reinforcement Learning Problems Through Partial Policy Recycling

Jan Ramon; Kurt Driessens; Tom Croonenborghs

We investigate the relation between transfer learning in reinforcement learning with function approximation and supervised learning with concept drift. We present a new incremental relational regression tree algorithm that is capable of dealing with concept drift through tree restructuring and show that it enables a Q-learner to transfer knowledge from one task to another by recycling those parts of the generalized Q-function that still hold interesting information for the new task. We illustrate the performance of the algorithm in several experiments.


pacific-asia conference on knowledge discovery and data mining | 2006

Using Weighted Nearest Neighbor to Benefit from Unlabeled Data.

Kurt Driessens; Peter Reutemann; Bernhard Pfahringer; Claire Leschi

The development of data-mining applications such as textclassification and molecular profiling has shown the need for machine learning algorithms that can benefit from both labeled and unlabeled data, where often the unlabeled examples greatly outnumber the labeled examples. In this paper we present a two-stage classifier that improves its predictive accuracy by making use of the available unlabeled data. It uses a weighted nearest neighbor classification algorithm using the combined example-sets as a knowledge base. The examples from the unlabeled set are “pre-labeled” by an initial classifier that is build using the limited available training data. By choosing appropriate weights for this pre-labeled data, the nearest neighbor classifier consistently improves on the original classifier.


Pattern Recognition Letters | 2015

Factored four way conditional restricted Boltzmann machines for activity recognition

Decebal Constantin Mocanu; Haitham Bou Ammar; Dietwig Jos Clement Lowet; Kurt Driessens; Antonio Liotta; Gerhard Weiss; Karl Tuyls

This paper proposes a new learning algorithm for human activity recognition.Its name is factored four way conditional restricted Boltzmann machine (FFW-CRBM).FFW-CRBMs are capable of simultaneous regression and classification.FFW-CRBMs came together with their own training procedure.The training procedure name is sequential Markov chain contrastive divergence. This paper introduces a new learning algorithm for human activity recognition capable of simultaneous regression and classification. Building upon conditional restricted Boltzmann machines (CRBMs), Factored four way conditional restricted Boltzmann machines (FFW-CRBMs) incorporate a new label layer and four-way interactions among the neurons from the different layers. The additional layer gives the classification nodes a similar strong multiplicative effect compared to the other layers, and avoids that the classification neurons are overwhelmed by the (much larger set of) other neurons. This makes FFW-CRBMs capable of performing activity recognition, prediction and self auto evaluation of classification within one unified framework. As a second contribution, sequential Markov chain contrastive divergence (SMcCD) is introduced. SMcCD modifies Contrastive Divergence to compensate for the extra complexity of FFW-CRBMs during training. Two sets of experiments one on benchmark datasets and one a robotic platform for smart companions show the effectiveness of FFW-CRBMs.


international conference on machine learning | 2005

Combining model-based and instance-based learning for first order regression

Kurt Driessens; Sašo Džeroski

The introduction of relational reinforcement learning and the RRL algorithm gave rise to the development of several first order regression algorithms. So far, these algorithms have employed either a model-based approach or an instance-based approach. As a consequence, they suffer from the typical drawbacks of model-based learning such as coarse function approximation or those of lazy learning such as high computational intensity.In this paper we develop a new regression algorithm that combines the strong points of both approaches and tries to avoid the normally inherent draw-backs. By combining model-based and instance-based learning, we produce an incremental first order regression algorithm that is both computationally efficient and produces better predictions earlier in the learning experiment.


european conference on machine learning | 2013

Automatically Mapped Transfer between Reinforcement Learning Tasks via Three-Way Restricted Boltzmann Machines

Haitham Bou Ammar; Decebal Constantin Mocanu; Matthew E. Taylor; Kurt Driessens; Karl Tuyls; Gerhard Weiss

Existing reinforcement learning approaches are often hampered by learning tabula rasa. Transfer for reinforcement learning tackles this problem by enabling the reuse of previously learned results, but may require an inter-task mapping to encode how the previously learned task and the new task are related. This paper presents an autonomous framework for learning inter-task mappings based on an adaptation of restricted Boltzmann machines. Both a full model and a computationally efficient factored model are introduced and shown to be effective in multiple transfer learning scenarios.


international conference on data mining | 2008

Cost-Sensitive Parsimonious Linear Regression

Robby Goetschalckx; Kurt Driessens; Scott Sanner

We examine linear regression problems where some features may only be observable at a cost (e.g., in medical domains where features may correspond to diagnostic tests that take time and costs money). This can be important in the context of data mining, in order to obtain the best predictions from the data on a limited cost budget. We define a parsimonious linear regression objective criterion that jointly minimizes prediction error and feature cost. We modify least angle regression algorithms commonly used for sparse linear regression to produce the ParLiR algorithm, which not only provides an efficient and parsimonious solution as we demonstrate empirically, but it also provides formal guarantees that we prove theoretically.

Collaboration


Dive into the Kurt Driessens's collaboration.

Top Co-Authors

Avatar

Jan Ramon

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar

Karl Tuyls

University of Liverpool

View shared research outputs
Top Co-Authors

Avatar

Tom Croonenborghs

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Haitham Bou Ammar

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Matthew E. Taylor

Washington State University

View shared research outputs
Top Co-Authors

Avatar

Hendrik Blockeel

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar

Nico Jacobs

Katholieke Universiteit Leuven

View shared research outputs
Top Co-Authors

Avatar

Robby Goetschalckx

Katholieke Universiteit Leuven

View shared research outputs
Researchain Logo
Decentralizing Knowledge