Pierrick Plamondon
Laval University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Pierrick Plamondon.
adaptive agents and multi-agents systems | 2007
Pierrick Plamondon; Brahim Chaib-draa; Abder Rezak Benaskeur
This paper contributes to solve effectively stochastic resource allocation problems known to be NP-Complete. To address this complex resource management problem, a Q-decomposition approach is proposed when the resources which are already shared among the agents, but the actions made by an agent may influence the reward obtained by at least another agent. The Q-decomposition allows to coordinate these reward separated agents and thus permits to reduce the set of states and actions to consider. On the other hand, when the resources are available to all agents, no Q-decomposition is possible and we use heuristic search. In particular, the bounded Real-time Dynamic Programming (bounded RTDP) is used. Bounded RTDP concentrates the planning on significant states only and prunes the action space. The pruning is accomplished by proposing tight upper and lower bounds on the value function.
international conference on knowledge-based and intelligent information and engineering systems | 2003
Pierrick Plamondon; Brahim Chaib-draa; Patrick Beaumont; Dale E. Blodgett
The position of a frigate to face some threats can augment its survival chances and therefore it is important to investigate this aspect in order to determine how a frigate can position itself during an attack. To achieve that, we propose a first method based on the Bayesian movement, performed by a learning agent, which determines the optimal positioning of the frigate by dividing the defense area into six sectors for weapon engagement and then, it makes efficient use of all the weapons available by using the sectors. The second method that we propose is called Radar Cross-Section Reduction (RCSR) movement and, it aims at reducing the exposed surface of the frigate to incoming threats before their locking phase is over. Preliminary results on these two methods are presented and discussed. Finally, an implementation of a meta-level agent which would make efficient use of both complementary methods is suggested.
ieee/wic/acm international conference on intelligent agent technology | 2005
Pierrick Plamondon; Brahim Chaib-draa; Abder Rezak Benaskeur
We are interested by contributing to stochastic problems of which the main distinction is that some tasks may create other tasks. In particular, we present a first approach which represents the problem by an acyclic graph, and solves each node in a certain order so as to produce an optimal solution. Then, we detail a second algorithm, which solves each task separately, using the first approach, and where an on-line heuristic computes the global actions to execute when the state of a task changes.
canadian conference on artificial intelligence | 2007
Camille Besse; Pierrick Plamondon; Brahim Chaib-draa
Resource allocation is a widely studied class of problems in Operation Research and Artificial Intelligence. Specially, constrained stochastic resource allocation problems, where the assignment of a constrained resource do not automatically imply the realization of the task. This kind of problems are generally addressed with Markov Decision Processes ( mdp s). In this paper, we present efficient lower and upper bounds in the context of a constrained stochastic resource allocation problem for a heuristic search algorithm called Focused Real Time Dynamic Programming ( frtdp ). Experiments show that this algorithm is relevant for this kind of problems and that the proposed tight bounds reduce the number of backups to perform comparatively to previous existing bounds.
International Journal on Artificial Intelligence Tools | 2012
Pierrick Plamondon; Brahim Chaib-draa
This paper contributes to solve effectively stochastic resource allocation problems in multiagent environments. To address it, a distributed Q-values approach is proposed when the resources are dis...
conference on decision and control | 2007
Pierrick Plamondon; Brahim Chaib-draa; Abder Rezak Benaskeur
This paper contributes to solve effectively stochastic resource allocation problems known to be NP-complete. To address this complex resource management problem, the merging of two approaches is made: The Q-decomposition model, which coordinates reward separated agents through an arbitrator, and the Labeled Real-Time Dynamic Programming (LRTDP) approaches are adapted in an effective way. The Q-decomposition permits to reduce the set of states to consider, while LRTDP concentrates the planning on significant states only. As demonstrated by the experiments, combining these two distinct approaches permits to further reduce the planning time to obtain the optimal solution of a resource allocation problem.
ieee wic acm international conference on intelligent agent technology | 2006
Pierrick Plamondon; Brahim Chaib-draa; Abder Rezak Benaskeur
This paper contributes to solve effectively stochastic resource allocation problems known to be NP-complete. To address this complex resource management problem, the merging of two approaches is made: The Q-decomposition model, which coordinates reward separated agents through an arbitrator, and the labeled real-time dynamic programming (LRTDP) approaches are adapted in an effective way. The Q-decomposition permits to reduce the set of states to consider, while LRTDP concentrates the planning on significant states only. As demonstrated by the experiments, combining these two distinct approaches permits to further reduce the planning time to obtain the optimal solution of a resource allocation problem.
canadian conference on artificial intelligence | 2006
Pierrick Plamondon; Brahim Chaib-draa; Abder Rezak Benaskeur
We are interested in contributing to solving effectively a particular type of real-time stochastic resource allocation problem. Firstly, one distinction is that certain tasks may create other tasks. Then, positive and negative interactions among the resources are considered, in achieving the tasks, in order to obtain and maintain an efficient coordination. A standard Multiagent Markov Decision Process (MMDP) approach is too prohibitive to solve this type of problem in real-time. To address this complex resource management problem, the merging of an approach which considers the complexity associated to a high number of different resource types (i.e. Multiagent Task Associated Markov Decision Processes (MTAMDP)), with an approach which considers the complexity associated to the creation of task by other tasks (i.e. Acyclic Decomposition) is proposed. The combination of these two approaches produces a near-optimal solution in much less time than a standard MMDP approach.
7th International Command and Control Research Technology Symposium | 2002
Dale E. Blodgett; Brahim Chaib-draaa; Pierrick Plamondon; Peter Kropf; Eloi Bosse
Intent Inference for Collaborative Tasks | 2001
Dale E. Blodgett; Sébastien Paquet; Pierrick Plamondon; Brahim Chaib-draa; Peter Kropf