Control Engineering Practice | 2021

Adaptive dynamic programming and deep reinforcement learning for the control of an unmanned surface vehicle: Experimental results

 
 
 
 

Abstract


Abstract This paper presents a low-level controller for an unmanned surface vehicle based on adaptive dynamic programming and deep reinforcement learning. This approach uses a single deep neural network capable of self-learning a policy, and driving the surge speed and yaw dynamics of a vessel. A simulation of the vehicle mathematical model was used to train the neural network with the model-based backpropagation through time algorithm, capable of dealing with continuous action-spaces. The path-following control scenario is additionally addressed by combining the proposed low-level controller and a line-of-sight based guidance law with time-varying look-ahead distance. Simulation and real-world experimental results are presented to validate the control capabilities of the proposed approach and contribute to the diversity of validated applications of adaptive dynamic programming based control strategies. Results show the controller is capable of self-learning the policy to drive the surge speed and yaw dynamics, and has an improved performance in comparison to a standard controller.

Volume 111
Pages 104807
DOI 10.1016/J.CONENGPRAC.2021.104807
Language English
Journal Control Engineering Practice

Full Text