Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Patrick Walters is active.

Publication


Featured researches published by Patrick Walters.


IEEE Transactions on Robotics | 2014

Nonlinear RISE-Based Control of an Autonomous Underwater Vehicle

Nicholas R. Fischer; Devin Hughes; Patrick Walters; Eric M. Schwartz; Warren E. Dixon

This study focuses on the development of a nonlinear control design for a fully-actuated autonomous underwater vehicle (AUV) using a continuous robust integral of the sign of the error control structure to compensate for system uncertainties and sufficiently smooth bounded exogenous disturbances. A Lyapunov stability analysis is included to prove semiglobal asymptotic tracking. The resulting controller is experimentally validated on an AUV developed at the University of Florida in both controlled and open-water environments.


Automatica | 2016

Model-based reinforcement learning for approximate optimal regulation

Rushikesh Kamalapurkar; Patrick Walters; Warren E. Dixon

Reinforcement learning (RL)-based online approximate optimal control methods applied to deterministic systems typically require a restrictive persistence of excitation (PE) condition for convergence. This paper develops a concurrent learning (CL)-based implementation of model-based RL to solve approximate optimal regulation problems online under a PE-like rank condition. The development is based on the observation that, given a model of the system, RL can be implemented by evaluating the Bellman error at any number of desired points in the state space. In this result, a parametric system model is considered, and a CL-based parameter identifier is developed to compensate for uncertainty in the parameters. Uniformly ultimately bounded regulation of the system states to a neighborhood of the origin, and convergence of the developed policy to a neighborhood of the optimal policy are established using a Lyapunov-based analysis, and simulation results are presented to demonstrate the performance of the developed controller.


conference on decision and control | 2013

Concurrent learning-based approximate optimal regulation

Rushikesh Kamalapurkar; Patrick Walters; Warren E. Dixon

In deterministic systems, reinforcement learning-based online approximate optimal control methods typically require a restrictive persistence of excitation (PE) condition for convergence. This paper presents a concurrent learning-based solution to the online approximate optimal regulation problem that eliminates the need for PE. The development is based on the observation that given a model of the system, the Bellman error, which quantifies the deviation of the system Hamiltonian from the optimal Hamiltonian, can be evaluated at any point in the state space. Further, a concurrent learning-based parameter identifier is developed to compensate for parametric uncertainty in the plant dynamics. Uniformly ultimately bounded (UUB) convergence of the system states to the origin, and UUB convergence of the developed policy to an approximate optimal policy are established using a Lyapunov-based analysis, and simulations are performed to demonstrate the performance of the developed controller.


conference on decision and control | 2014

Model-based reinforcement learning for infinite-horizon approximate optimal tracking

Rushikesh Kamalapurkar; Lindsey Andrews; Patrick Walters; Warren E. Dixon

This brief paper provides an approximate online adaptive solution to the infinite-horizon optimal tracking problem for control-affine continuous-time nonlinear systems with unknown drift dynamics. To relax the persistence of excitation condition, model-based reinforcement learning is implemented using a concurrent-learning-based system identifier to simulate experience by evaluating the Bellman error over unexplored areas of the state space. Tracking of the desired trajectory and convergence of the developed policy to a neighborhood of the optimal policy are established via Lyapunov-based stability analysis. Simulation results demonstrate the effectiveness of the developed technique.


IEEE Transactions on Control of Network Systems | 2018

Model-Based Reinforcement Learning in Differential Graphical Games

Rushikesh Kamalapurkar; Justin R. Klotz; Patrick Walters; Warren E. Dixon

This paper seeks to combine differential game theory with the actor-critic-identifier architecture to determine forward-in-time, approximate optimal controllers for formation tracking in multiagent systems, where the agents have uncertain heterogeneous nonlinear dynamics. A continuous control strategy is proposed, using communication feedback from extended neighbors on a communication topology that has a spanning tree. A model-based reinforcement learning technique is developed to cooperatively control a group of agents to track a trajectory in a desired formation. Simulation results are presented to demonstrate the performance of the developed technique.


conference on decision and control | 2014

Online approximate optimal path-following for a mobile robot

Patrick Walters; Rushikesh Kamalapurkar; Lindsey Andrews; Warren E. Dixon

Online approximation of an infinite horizon optimal path-following strategy for a unicycle-type mobile robot is considered. An approximate solution to the optimal control problem is obtained by using an adaptive dynamic programming technique that uses adaptive update laws to estimate the unknown value function. The developed controller overcomes challenges with the approximation of the infinite horizon value function by using an auxiliary function that describes the motion of a virtual target on the desired path. The developed controller guarantees uniformly ultimately bounded (UUB) convergence of the vehicle to a desired path while maintaining a desired speed profile and UUB convergence of the approximate policy to the optimal policy without requiring persistence of excitation.


Archive | 2018

Model-Based Reinforcement Learning for Approximate Optimal Control

Rushikesh Kamalapurkar; Patrick Walters; Joel A. Rosenfeld; Warren E. Dixon

This chapter develops a data-driven implementation of model-based reinforcement learning to solve approximate optimal control problems online under a persistence of excitation-like rank condition. The development is based on the observation that, given a model of the system, reinforcement learning can be implemented by evaluating the Bellman error at any number of desired points in the state-space. In this result, a parametric system model is considered, and a data-driven parameter identifier is developed to compensate for uncertainty in the parameters. Uniformly ultimately bounded regulation of the system states to a neighborhood of the origin, and convergence of the developed policy to a neighborhood of the optimal policy are established using a Lyapunov-based analysis. Simulation results indicate that the developed controller can be implemented to achieve fast online learning without the addition of ad-hoc probing signals as in Chap. 3. The developed model-based reinforcement learning method is extended to solve trajectory tracking problems for uncertain nonlinear systems, and to generate approximate feedback-Nash equilibrium solutions to N-player nonzero-sum differential games.


Archive | 2018

Differential Graphical Games

Rushikesh Kamalapurkar; Patrick Walters; Joel A. Rosenfeld; Warren E. Dixon

This chapter deals with the formulation and online approximate feedback-Nash equilibrium solution of an optimal network formation tracking problem. A relative control error minimization technique is introduced to facilitate the formulation of a feasible infinite-horizon total-cost differential graphical game. A dynamic programming-based feedback-Nash equilibrium solution to the differential graphical game is obtained via the development of a set of coupled Hamilton–Jacobi equations. The developed approximate feedback-Nash equilibrium solution is analyzed using a Lyapunov-based stability analysis to demonstrate ultimately bounded formation tracking in the presence of uncertainties. In addition to control, this chapter also explores applications of differential graphical games to monitoring the behavior of neighboring agents in a network.


Archive | 2018

Approximate Dynamic Programming

Rushikesh Kamalapurkar; Patrick Walters; Joel A. Rosenfeld; Warren E. Dixon

This chapter contains a brief review of dynamic programming in continuous time and space. In particular, traditional dynamic programming algorithms such as policy iteration, value iteration, and actor-critic methods are presented in the context of continuous-time optimal control. The role of the optimal value function as a Lyapunov function is explained to facilitate online closed-loop optimal control. This chapter also highlights the problems and the limitations of existing techniques, thereby motivating the development in this book. The chapter concludes with some historic remarks and a brief classification of the available dynamic programming techniques.


Archive | 2018

Excitation-Based Online Approximate Optimal Control

Rushikesh Kamalapurkar; Patrick Walters; Joel A. Rosenfeld; Warren E. Dixon

In this chapter, online adaptive reinforcement learning-based solutions are developed for infinite-horizon optimal control problems for continuous-time uncertain nonlinear systems. An actor-critic-identifier structure is developed to approximate the solution to the Hamilton–Jacobi–Bellman equation using three neural network structures. The actor and the critic neural networks approximate the optimal control and the optimal value function, respectively, and a robust dynamic neural network identifier asymptotically approximates the uncertain system dynamics. An advantage of the using the actor-critic-identifier architecture is that learning by the actor, critic, and identifier is continuous and concurrent, without requiring knowledge of system drift dynamics. Convergence of the algorithm is analyzed using Lyapunov-based adaptive control methods. A persistence of excitation condition is required to guarantee exponential convergence to a bounded region in the neighborhood of the optimal control and uniformly ultimately bounded stability of the closed-loop system. The developed actor-critic method is extended to solve trajectory tracking problems under the assumption that the system dynamics are completely known. The actor-critic-identifier architecture is also extended to generate approximate feedback-Nash equilibrium solutions to N-player nonzero-sum differential games. Simulation results are provided to demonstrate the performance of the developed actor-critic-identifier method.

Collaboration


Dive into the Patrick Walters's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge